var/home/core/zuul-output/0000755000175000017500000000000015115032476014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015115036575015501 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003767226315115036566017725 0ustar rootrootDec 06 13:55:56 crc systemd[1]: Starting Kubernetes Kubelet... Dec 06 13:55:56 crc restorecon[4691]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:56 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 13:55:57 crc restorecon[4691]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 13:55:57 crc restorecon[4691]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 06 13:55:57 crc kubenswrapper[4706]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 13:55:57 crc kubenswrapper[4706]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 06 13:55:57 crc kubenswrapper[4706]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 13:55:57 crc kubenswrapper[4706]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 13:55:57 crc kubenswrapper[4706]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 06 13:55:57 crc kubenswrapper[4706]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.279098 4706 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282578 4706 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282598 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282604 4706 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282609 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282614 4706 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282618 4706 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282624 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282628 4706 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282635 4706 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282646 4706 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282651 4706 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282657 4706 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282661 4706 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282664 4706 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282668 4706 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282672 4706 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282676 4706 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282680 4706 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282684 4706 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282688 4706 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282692 4706 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282695 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282699 4706 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282704 4706 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282708 4706 feature_gate.go:330] unrecognized feature gate: Example Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282711 4706 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282715 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282720 4706 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282723 4706 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282727 4706 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282730 4706 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282734 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282737 4706 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282741 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282744 4706 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282748 4706 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282752 4706 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282756 4706 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282759 4706 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282763 4706 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282766 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282770 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282775 4706 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282778 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282782 4706 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282785 4706 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282788 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282792 4706 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282795 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282799 4706 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282802 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282805 4706 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282809 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282814 4706 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282818 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282822 4706 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282826 4706 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282829 4706 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282833 4706 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282836 4706 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282840 4706 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282843 4706 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282849 4706 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282853 4706 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282857 4706 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282861 4706 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282865 4706 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282869 4706 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282873 4706 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282877 4706 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.282882 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.282950 4706 flags.go:64] FLAG: --address="0.0.0.0" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.282958 4706 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.282964 4706 flags.go:64] FLAG: --anonymous-auth="true" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.282970 4706 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.282976 4706 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.282980 4706 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.282986 4706 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.282991 4706 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.282996 4706 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283000 4706 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283005 4706 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283009 4706 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283013 4706 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283017 4706 flags.go:64] FLAG: --cgroup-root="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283021 4706 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283025 4706 flags.go:64] FLAG: --client-ca-file="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283029 4706 flags.go:64] FLAG: --cloud-config="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283034 4706 flags.go:64] FLAG: --cloud-provider="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283038 4706 flags.go:64] FLAG: --cluster-dns="[]" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283044 4706 flags.go:64] FLAG: --cluster-domain="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283049 4706 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283054 4706 flags.go:64] FLAG: --config-dir="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283058 4706 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283063 4706 flags.go:64] FLAG: --container-log-max-files="5" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283069 4706 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283074 4706 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283078 4706 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283084 4706 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283089 4706 flags.go:64] FLAG: --contention-profiling="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283094 4706 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283099 4706 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283105 4706 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283111 4706 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283117 4706 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283122 4706 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283127 4706 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283132 4706 flags.go:64] FLAG: --enable-load-reader="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283137 4706 flags.go:64] FLAG: --enable-server="true" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283142 4706 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283148 4706 flags.go:64] FLAG: --event-burst="100" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283153 4706 flags.go:64] FLAG: --event-qps="50" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283157 4706 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283161 4706 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283165 4706 flags.go:64] FLAG: --eviction-hard="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283170 4706 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283174 4706 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283179 4706 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283183 4706 flags.go:64] FLAG: --eviction-soft="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283188 4706 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283192 4706 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283196 4706 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283200 4706 flags.go:64] FLAG: --experimental-mounter-path="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283204 4706 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283208 4706 flags.go:64] FLAG: --fail-swap-on="true" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283212 4706 flags.go:64] FLAG: --feature-gates="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283217 4706 flags.go:64] FLAG: --file-check-frequency="20s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283221 4706 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283225 4706 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283230 4706 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283234 4706 flags.go:64] FLAG: --healthz-port="10248" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283238 4706 flags.go:64] FLAG: --help="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283242 4706 flags.go:64] FLAG: --hostname-override="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283246 4706 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283251 4706 flags.go:64] FLAG: --http-check-frequency="20s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283255 4706 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283259 4706 flags.go:64] FLAG: --image-credential-provider-config="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283263 4706 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283267 4706 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283271 4706 flags.go:64] FLAG: --image-service-endpoint="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283275 4706 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283279 4706 flags.go:64] FLAG: --kube-api-burst="100" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283284 4706 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283288 4706 flags.go:64] FLAG: --kube-api-qps="50" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283292 4706 flags.go:64] FLAG: --kube-reserved="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283297 4706 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283301 4706 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283321 4706 flags.go:64] FLAG: --kubelet-cgroups="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283325 4706 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283329 4706 flags.go:64] FLAG: --lock-file="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283334 4706 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283338 4706 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283343 4706 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283349 4706 flags.go:64] FLAG: --log-json-split-stream="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283353 4706 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283357 4706 flags.go:64] FLAG: --log-text-split-stream="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283361 4706 flags.go:64] FLAG: --logging-format="text" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283365 4706 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283369 4706 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283373 4706 flags.go:64] FLAG: --manifest-url="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283377 4706 flags.go:64] FLAG: --manifest-url-header="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283382 4706 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283387 4706 flags.go:64] FLAG: --max-open-files="1000000" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283392 4706 flags.go:64] FLAG: --max-pods="110" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283396 4706 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283400 4706 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283404 4706 flags.go:64] FLAG: --memory-manager-policy="None" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283408 4706 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283412 4706 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283416 4706 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283421 4706 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283430 4706 flags.go:64] FLAG: --node-status-max-images="50" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283434 4706 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283439 4706 flags.go:64] FLAG: --oom-score-adj="-999" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283443 4706 flags.go:64] FLAG: --pod-cidr="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283446 4706 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283452 4706 flags.go:64] FLAG: --pod-manifest-path="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283456 4706 flags.go:64] FLAG: --pod-max-pids="-1" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283461 4706 flags.go:64] FLAG: --pods-per-core="0" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283465 4706 flags.go:64] FLAG: --port="10250" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283469 4706 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283473 4706 flags.go:64] FLAG: --provider-id="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283477 4706 flags.go:64] FLAG: --qos-reserved="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283481 4706 flags.go:64] FLAG: --read-only-port="10255" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283485 4706 flags.go:64] FLAG: --register-node="true" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283489 4706 flags.go:64] FLAG: --register-schedulable="true" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283493 4706 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283500 4706 flags.go:64] FLAG: --registry-burst="10" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283505 4706 flags.go:64] FLAG: --registry-qps="5" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283509 4706 flags.go:64] FLAG: --reserved-cpus="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283512 4706 flags.go:64] FLAG: --reserved-memory="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283518 4706 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283522 4706 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283526 4706 flags.go:64] FLAG: --rotate-certificates="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283530 4706 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283534 4706 flags.go:64] FLAG: --runonce="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283538 4706 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283542 4706 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283546 4706 flags.go:64] FLAG: --seccomp-default="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283550 4706 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283554 4706 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283558 4706 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283562 4706 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283568 4706 flags.go:64] FLAG: --storage-driver-password="root" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283573 4706 flags.go:64] FLAG: --storage-driver-secure="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283577 4706 flags.go:64] FLAG: --storage-driver-table="stats" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283581 4706 flags.go:64] FLAG: --storage-driver-user="root" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283585 4706 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283590 4706 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283594 4706 flags.go:64] FLAG: --system-cgroups="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283598 4706 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283604 4706 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283610 4706 flags.go:64] FLAG: --tls-cert-file="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283614 4706 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283619 4706 flags.go:64] FLAG: --tls-min-version="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283623 4706 flags.go:64] FLAG: --tls-private-key-file="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283627 4706 flags.go:64] FLAG: --topology-manager-policy="none" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283632 4706 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283636 4706 flags.go:64] FLAG: --topology-manager-scope="container" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283640 4706 flags.go:64] FLAG: --v="2" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283645 4706 flags.go:64] FLAG: --version="false" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283650 4706 flags.go:64] FLAG: --vmodule="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283655 4706 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.283659 4706 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283748 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283753 4706 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283758 4706 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283763 4706 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283767 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283771 4706 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283775 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283780 4706 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283784 4706 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283788 4706 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283792 4706 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283798 4706 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283803 4706 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283807 4706 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283810 4706 feature_gate.go:330] unrecognized feature gate: Example Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283814 4706 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283818 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283822 4706 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283826 4706 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283829 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283834 4706 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283838 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283842 4706 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283845 4706 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283849 4706 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283852 4706 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283856 4706 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283859 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283863 4706 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283866 4706 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283870 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283873 4706 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283876 4706 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283881 4706 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283886 4706 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283890 4706 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283894 4706 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283898 4706 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283902 4706 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283906 4706 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283910 4706 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283913 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283917 4706 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283922 4706 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283926 4706 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283929 4706 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283933 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283937 4706 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283940 4706 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283944 4706 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283947 4706 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283951 4706 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283956 4706 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283959 4706 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283963 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283967 4706 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283970 4706 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283974 4706 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283978 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283981 4706 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283985 4706 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283988 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283991 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283995 4706 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.283998 4706 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.284002 4706 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.284005 4706 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.284009 4706 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.284012 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.284016 4706 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.284019 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.284025 4706 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.294561 4706 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.294610 4706 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294788 4706 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294802 4706 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294813 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294823 4706 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294834 4706 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294848 4706 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294858 4706 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294866 4706 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294875 4706 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294882 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294890 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294898 4706 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294906 4706 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294914 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294922 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294929 4706 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294938 4706 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294946 4706 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294955 4706 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294963 4706 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294971 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294978 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294987 4706 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.294995 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295003 4706 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295010 4706 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295019 4706 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295027 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295035 4706 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295042 4706 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295050 4706 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295059 4706 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295067 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295075 4706 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295082 4706 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295090 4706 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295098 4706 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295106 4706 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295113 4706 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295121 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295129 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295137 4706 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295145 4706 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295153 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295161 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295169 4706 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295177 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295184 4706 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295193 4706 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295204 4706 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295213 4706 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295222 4706 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295231 4706 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295239 4706 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295247 4706 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295254 4706 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295262 4706 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295269 4706 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295280 4706 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295289 4706 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295298 4706 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295331 4706 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295341 4706 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295351 4706 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295362 4706 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295373 4706 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295383 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295392 4706 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295401 4706 feature_gate.go:330] unrecognized feature gate: Example Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295409 4706 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295417 4706 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.295431 4706 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295655 4706 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295668 4706 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295676 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295686 4706 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295695 4706 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295704 4706 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295712 4706 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295721 4706 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295733 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295743 4706 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295753 4706 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295763 4706 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295776 4706 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295792 4706 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295804 4706 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295816 4706 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295826 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295836 4706 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295845 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295853 4706 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295860 4706 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295868 4706 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295876 4706 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295884 4706 feature_gate.go:330] unrecognized feature gate: Example Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295895 4706 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295904 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295914 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295924 4706 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295933 4706 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295942 4706 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295950 4706 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295959 4706 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295968 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295976 4706 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295985 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.295993 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296001 4706 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296009 4706 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296016 4706 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296025 4706 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296035 4706 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296044 4706 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296052 4706 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296061 4706 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296069 4706 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296078 4706 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296086 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296094 4706 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296102 4706 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296109 4706 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296117 4706 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296125 4706 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296145 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296153 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296161 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296169 4706 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296178 4706 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296206 4706 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296214 4706 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296222 4706 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296231 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296238 4706 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296247 4706 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296255 4706 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296263 4706 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296271 4706 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296279 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296287 4706 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296294 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296302 4706 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.296338 4706 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.296352 4706 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.296624 4706 server.go:940] "Client rotation is on, will bootstrap in background" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.301576 4706 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.301640 4706 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.302167 4706 server.go:997] "Starting client certificate rotation" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.302186 4706 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.302776 4706 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-03 04:58:42.013524135 +0000 UTC Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.302906 4706 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 663h2m44.7106232s for next certificate rotation Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.306821 4706 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.308278 4706 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.363144 4706 log.go:25] "Validated CRI v1 runtime API" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.387811 4706 log.go:25] "Validated CRI v1 image API" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.391516 4706 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.394571 4706 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-06-13-52-07-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.394773 4706 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.420139 4706 manager.go:217] Machine: {Timestamp:2025-12-06 13:55:57.418052589 +0000 UTC m=+0.253845526 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:0e9ac240-47ff-4359-8abf-1e7458f9a542 BootID:6d0ebee1-8e8f-40f3-be2d-d81f84254664 Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:ab:25:8d Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:ab:25:8d Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:fe:b7:8b Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:cb:48:88 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:e3:42:de Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:60:fb:35 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:3a:26:b1:9a:79:89 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:d2:15:4e:a0:cd:fc Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.420798 4706 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.421080 4706 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.421916 4706 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.422365 4706 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.422558 4706 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.422996 4706 topology_manager.go:138] "Creating topology manager with none policy" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.423108 4706 container_manager_linux.go:303] "Creating device plugin manager" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.423549 4706 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.423720 4706 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.424378 4706 state_mem.go:36] "Initialized new in-memory state store" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.425092 4706 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.426493 4706 kubelet.go:418] "Attempting to sync node with API server" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.426891 4706 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.427464 4706 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.427872 4706 kubelet.go:324] "Adding apiserver pod source" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.428063 4706 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.433454 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Dec 06 13:55:57 crc kubenswrapper[4706]: E1206 13:55:57.433626 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.433705 4706 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.433542 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Dec 06 13:55:57 crc kubenswrapper[4706]: E1206 13:55:57.433868 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.434250 4706 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.435512 4706 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.436440 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.436481 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.436495 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.436509 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.436547 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.436575 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.436593 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.436623 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.436639 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.436654 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.436673 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.436689 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.436957 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.437780 4706 server.go:1280] "Started kubelet" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.438214 4706 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.438366 4706 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.438363 4706 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.439497 4706 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 06 13:55:57 crc systemd[1]: Started Kubernetes Kubelet. Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.440254 4706 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.440370 4706 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 06 13:55:57 crc kubenswrapper[4706]: E1206 13:55:57.441362 4706 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.441430 4706 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.441459 4706 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.441498 4706 server.go:460] "Adding debug handlers to kubelet server" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.441634 4706 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.440403 4706 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 16:12:55.479407435 +0000 UTC Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.441804 4706 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 530h16m58.037624376s for next certificate rotation Dec 06 13:55:57 crc kubenswrapper[4706]: E1206 13:55:57.442233 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="200ms" Dec 06 13:55:57 crc kubenswrapper[4706]: E1206 13:55:57.440028 4706 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.83:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187ea4d95c365bbd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 13:55:57.437737917 +0000 UTC m=+0.273530804,LastTimestamp:2025-12-06 13:55:57.437737917 +0000 UTC m=+0.273530804,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.443616 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Dec 06 13:55:57 crc kubenswrapper[4706]: E1206 13:55:57.443731 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.447869 4706 factory.go:55] Registering systemd factory Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.447924 4706 factory.go:221] Registration of the systemd container factory successfully Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.448487 4706 factory.go:153] Registering CRI-O factory Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.448520 4706 factory.go:221] Registration of the crio container factory successfully Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.448639 4706 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.448685 4706 factory.go:103] Registering Raw factory Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.448712 4706 manager.go:1196] Started watching for new ooms in manager Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.449909 4706 manager.go:319] Starting recovery of all containers Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465155 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465248 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465273 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465296 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465345 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465370 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465394 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465416 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465439 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465460 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465482 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465503 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465523 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465548 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465572 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465592 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465617 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465636 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465656 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465678 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465704 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465726 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465749 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465771 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465792 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465852 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465880 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465911 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.465945 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466013 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466043 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466066 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466089 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466111 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466162 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466185 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466210 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466233 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466255 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466278 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466298 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466346 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466371 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466394 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466419 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466442 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466468 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466492 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466511 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466532 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466554 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466576 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466605 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466630 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466655 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466680 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466702 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466722 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466742 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466761 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466783 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466802 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466824 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466845 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466865 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466885 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466904 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466925 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466947 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466967 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.466984 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467002 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467022 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467042 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467064 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467084 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467107 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467127 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467148 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467168 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467188 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467208 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467231 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467254 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467277 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467298 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467384 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467408 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467429 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467451 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467477 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467498 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467518 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467539 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467560 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467583 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467606 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467630 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467649 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467671 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467691 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467712 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467731 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467753 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467782 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467803 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467825 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467849 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467871 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467896 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467920 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467941 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467962 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.467984 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468004 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468027 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468053 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468082 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468107 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468134 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468164 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468189 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468210 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468231 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468256 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468305 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468358 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468380 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468400 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468420 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468439 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468459 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468481 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468502 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468522 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468543 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468562 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468584 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468604 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468622 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468641 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468662 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468687 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468706 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468726 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468747 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468768 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468789 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468814 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468835 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468858 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468879 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468900 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468925 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468945 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468967 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.468990 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.469025 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.469048 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.469069 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.469091 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.469116 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.469139 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.469163 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.469186 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.469209 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.469232 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.472539 4706 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.472834 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.472878 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.472912 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.472941 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.472971 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.472999 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473036 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473069 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473101 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473134 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473162 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473189 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473218 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473245 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473272 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473303 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473390 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473417 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473455 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473487 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473514 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473540 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473569 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473597 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473624 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473662 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473692 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473718 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473748 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473776 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.473976 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.474009 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.474037 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.474064 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.474095 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.474124 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.474150 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.474176 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.474205 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.474230 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.474254 4706 reconstruct.go:97] "Volume reconstruction finished" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.474272 4706 reconciler.go:26] "Reconciler: start to sync state" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.486463 4706 manager.go:324] Recovery completed Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.501237 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.502679 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.502708 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.502717 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.503525 4706 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.503539 4706 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.503555 4706 state_mem.go:36] "Initialized new in-memory state store" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.506855 4706 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.509728 4706 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.509786 4706 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.509826 4706 kubelet.go:2335] "Starting kubelet main sync loop" Dec 06 13:55:57 crc kubenswrapper[4706]: E1206 13:55:57.509888 4706 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.511771 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Dec 06 13:55:57 crc kubenswrapper[4706]: E1206 13:55:57.511995 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.512984 4706 policy_none.go:49] "None policy: Start" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.514108 4706 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.514133 4706 state_mem.go:35] "Initializing new in-memory state store" Dec 06 13:55:57 crc kubenswrapper[4706]: E1206 13:55:57.541672 4706 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.589793 4706 manager.go:334] "Starting Device Plugin manager" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.589834 4706 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.589845 4706 server.go:79] "Starting device plugin registration server" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.590176 4706 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.590192 4706 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.590411 4706 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.590587 4706 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.590602 4706 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 06 13:55:57 crc kubenswrapper[4706]: E1206 13:55:57.600302 4706 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.610045 4706 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.610184 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.611748 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.611816 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.611834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.612274 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.612455 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.612512 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.613797 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.613872 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.613934 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.614027 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.614074 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.614093 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.614230 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.614482 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.614557 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.615188 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.615224 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.615241 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.615402 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.615690 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.615726 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.615740 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.615964 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.616001 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.616895 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.616924 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.616935 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.617000 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.617039 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.617056 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.617288 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.617427 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.617468 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.618354 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.618399 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.618415 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.618556 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.618591 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.618607 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.618837 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.618884 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.619990 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.620023 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.620035 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:57 crc kubenswrapper[4706]: E1206 13:55:57.642955 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="400ms" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.676587 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.676690 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.676759 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.676834 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.676941 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.677087 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.677195 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.677287 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.677437 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.677550 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.677647 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.677753 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.677862 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.677993 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.678084 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.692730 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.694480 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.694558 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.694589 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.694655 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 13:55:57 crc kubenswrapper[4706]: E1206 13:55:57.695237 4706 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.83:6443: connect: connection refused" node="crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.778834 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.779066 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.779240 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.779476 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.779704 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.779546 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.779289 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.779816 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.780216 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.780012 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.780688 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.780723 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.780785 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.780814 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.780876 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.780905 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.780966 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.780995 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.781057 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.781085 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.781397 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.781522 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.781587 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.781496 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.781503 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.781449 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.781526 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.781668 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.781478 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.782035 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.896404 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.898947 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.898991 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.899008 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.899043 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 13:55:57 crc kubenswrapper[4706]: E1206 13:55:57.899642 4706 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.83:6443: connect: connection refused" node="crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.952775 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.964751 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.984612 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-7e899cb783f52a59e3151d993cc8c355a0e0e8f7ddb35b0de143a8f1e41edc7c WatchSource:0}: Error finding container 7e899cb783f52a59e3151d993cc8c355a0e0e8f7ddb35b0de143a8f1e41edc7c: Status 404 returned error can't find the container with id 7e899cb783f52a59e3151d993cc8c355a0e0e8f7ddb35b0de143a8f1e41edc7c Dec 06 13:55:57 crc kubenswrapper[4706]: W1206 13:55:57.990394 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-b6bf9652b7348ea6f58e1e1256ee9b0d5eeb092cd6eea8bfe6f9b2a274131f78 WatchSource:0}: Error finding container b6bf9652b7348ea6f58e1e1256ee9b0d5eeb092cd6eea8bfe6f9b2a274131f78: Status 404 returned error can't find the container with id b6bf9652b7348ea6f58e1e1256ee9b0d5eeb092cd6eea8bfe6f9b2a274131f78 Dec 06 13:55:57 crc kubenswrapper[4706]: I1206 13:55:57.990947 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.016629 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.027039 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:55:58 crc kubenswrapper[4706]: E1206 13:55:58.044961 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="800ms" Dec 06 13:55:58 crc kubenswrapper[4706]: W1206 13:55:58.052896 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-35c022e471a8be39f2ca7767a20018c3fd5af3a4a2a7562f84c24f15972ec27b WatchSource:0}: Error finding container 35c022e471a8be39f2ca7767a20018c3fd5af3a4a2a7562f84c24f15972ec27b: Status 404 returned error can't find the container with id 35c022e471a8be39f2ca7767a20018c3fd5af3a4a2a7562f84c24f15972ec27b Dec 06 13:55:58 crc kubenswrapper[4706]: W1206 13:55:58.057302 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-539bc4440468819b21fdd2b2b223281bdc1fc0ad3b538b71163df30a1659d3a8 WatchSource:0}: Error finding container 539bc4440468819b21fdd2b2b223281bdc1fc0ad3b538b71163df30a1659d3a8: Status 404 returned error can't find the container with id 539bc4440468819b21fdd2b2b223281bdc1fc0ad3b538b71163df30a1659d3a8 Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.299730 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.300987 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.301037 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.301045 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.301063 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 13:55:58 crc kubenswrapper[4706]: E1206 13:55:58.301360 4706 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.83:6443: connect: connection refused" node="crc" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.439343 4706 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.515950 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01" exitCode=0 Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.516080 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01"} Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.516201 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"539bc4440468819b21fdd2b2b223281bdc1fc0ad3b538b71163df30a1659d3a8"} Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.516557 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.518856 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.518908 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.518931 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.520945 4706 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe" exitCode=0 Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.521019 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe"} Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.521050 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"35c022e471a8be39f2ca7767a20018c3fd5af3a4a2a7562f84c24f15972ec27b"} Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.521197 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.522100 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.523063 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.523099 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.523116 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.523599 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.523626 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.523637 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.525982 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="e950c97382f0fa9088ca0fe05a5b383f141331728b6a7d98b2a57bb6102ae218" exitCode=0 Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.526040 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"e950c97382f0fa9088ca0fe05a5b383f141331728b6a7d98b2a57bb6102ae218"} Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.526062 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"07a7e2f362d0421b524a74ff3fec94c63e68de7baba5a808302cb64c24a243b9"} Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.526114 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.526947 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.526968 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.526978 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.528739 4706 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607" exitCode=0 Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.528812 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607"} Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.528885 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b6bf9652b7348ea6f58e1e1256ee9b0d5eeb092cd6eea8bfe6f9b2a274131f78"} Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.529014 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.530081 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.530123 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.530135 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.532521 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5"} Dec 06 13:55:58 crc kubenswrapper[4706]: I1206 13:55:58.532548 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7e899cb783f52a59e3151d993cc8c355a0e0e8f7ddb35b0de143a8f1e41edc7c"} Dec 06 13:55:58 crc kubenswrapper[4706]: W1206 13:55:58.552956 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Dec 06 13:55:58 crc kubenswrapper[4706]: E1206 13:55:58.553056 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Dec 06 13:55:58 crc kubenswrapper[4706]: W1206 13:55:58.589199 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Dec 06 13:55:58 crc kubenswrapper[4706]: E1206 13:55:58.589334 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Dec 06 13:55:58 crc kubenswrapper[4706]: W1206 13:55:58.715175 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Dec 06 13:55:58 crc kubenswrapper[4706]: E1206 13:55:58.715296 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Dec 06 13:55:58 crc kubenswrapper[4706]: E1206 13:55:58.846115 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="1.6s" Dec 06 13:55:58 crc kubenswrapper[4706]: W1206 13:55:58.868694 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.83:6443: connect: connection refused Dec 06 13:55:58 crc kubenswrapper[4706]: E1206 13:55:58.868783 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.83:6443: connect: connection refused" logger="UnhandledError" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.103257 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.104420 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.104454 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.104465 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.104493 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.538991 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"5a3f1b130b3e40c9e1863b1af3ea9e41c19d07fe95a0c767612e3fa203c50be6"} Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.539214 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.542128 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.542173 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.542185 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.545281 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0c44a305126e75b14cf63cc84ae01d0d04203db32abb5f4434f1339a8e773fca"} Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.545428 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0730aa973ef51ef8c4d240fde0ef6135c76ac40b576a1afa846e8b3772c16120"} Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.545461 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"631b8b4aeac64fbdd48083db4545f8df35fa2e62abb03fcfc0d78ee39216be3c"} Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.545615 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.546853 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.546893 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.546906 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.554299 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75"} Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.554406 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318"} Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.554423 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391"} Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.554566 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.555857 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.555896 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.555908 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.560461 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42"} Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.560487 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05"} Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.560501 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0"} Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.560513 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788"} Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.565605 4706 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4" exitCode=0 Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.565680 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4"} Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.565853 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.566996 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.567024 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:55:59 crc kubenswrapper[4706]: I1206 13:55:59.567033 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:00 crc kubenswrapper[4706]: I1206 13:56:00.573610 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3"} Dec 06 13:56:00 crc kubenswrapper[4706]: I1206 13:56:00.573754 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:00 crc kubenswrapper[4706]: I1206 13:56:00.575136 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:00 crc kubenswrapper[4706]: I1206 13:56:00.575200 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:00 crc kubenswrapper[4706]: I1206 13:56:00.575222 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:00 crc kubenswrapper[4706]: I1206 13:56:00.577187 4706 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b" exitCode=0 Dec 06 13:56:00 crc kubenswrapper[4706]: I1206 13:56:00.577272 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b"} Dec 06 13:56:00 crc kubenswrapper[4706]: I1206 13:56:00.577393 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:00 crc kubenswrapper[4706]: I1206 13:56:00.577472 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:00 crc kubenswrapper[4706]: I1206 13:56:00.578686 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:00 crc kubenswrapper[4706]: I1206 13:56:00.578742 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:00 crc kubenswrapper[4706]: I1206 13:56:00.578745 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:00 crc kubenswrapper[4706]: I1206 13:56:00.578767 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:00 crc kubenswrapper[4706]: I1206 13:56:00.578784 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:00 crc kubenswrapper[4706]: I1206 13:56:00.578802 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:01 crc kubenswrapper[4706]: I1206 13:56:01.586611 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d"} Dec 06 13:56:01 crc kubenswrapper[4706]: I1206 13:56:01.586695 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b"} Dec 06 13:56:01 crc kubenswrapper[4706]: I1206 13:56:01.586729 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753"} Dec 06 13:56:01 crc kubenswrapper[4706]: I1206 13:56:01.586746 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:01 crc kubenswrapper[4706]: I1206 13:56:01.586785 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:56:01 crc kubenswrapper[4706]: I1206 13:56:01.588278 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:01 crc kubenswrapper[4706]: I1206 13:56:01.588355 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:01 crc kubenswrapper[4706]: I1206 13:56:01.588372 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:01 crc kubenswrapper[4706]: I1206 13:56:01.920405 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:56:01 crc kubenswrapper[4706]: I1206 13:56:01.945028 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 13:56:01 crc kubenswrapper[4706]: I1206 13:56:01.945238 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:01 crc kubenswrapper[4706]: I1206 13:56:01.947635 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:01 crc kubenswrapper[4706]: I1206 13:56:01.947698 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:01 crc kubenswrapper[4706]: I1206 13:56:01.947718 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:02 crc kubenswrapper[4706]: I1206 13:56:02.598215 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913"} Dec 06 13:56:02 crc kubenswrapper[4706]: I1206 13:56:02.598277 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9"} Dec 06 13:56:02 crc kubenswrapper[4706]: I1206 13:56:02.598366 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:02 crc kubenswrapper[4706]: I1206 13:56:02.598432 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:02 crc kubenswrapper[4706]: I1206 13:56:02.600010 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:02 crc kubenswrapper[4706]: I1206 13:56:02.600064 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:02 crc kubenswrapper[4706]: I1206 13:56:02.600089 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:02 crc kubenswrapper[4706]: I1206 13:56:02.600223 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:02 crc kubenswrapper[4706]: I1206 13:56:02.600290 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:02 crc kubenswrapper[4706]: I1206 13:56:02.600340 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:02 crc kubenswrapper[4706]: I1206 13:56:02.636301 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 13:56:02 crc kubenswrapper[4706]: I1206 13:56:02.636656 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:02 crc kubenswrapper[4706]: I1206 13:56:02.638663 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:02 crc kubenswrapper[4706]: I1206 13:56:02.638724 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:02 crc kubenswrapper[4706]: I1206 13:56:02.638744 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:03 crc kubenswrapper[4706]: I1206 13:56:03.601605 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:03 crc kubenswrapper[4706]: I1206 13:56:03.601670 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:03 crc kubenswrapper[4706]: I1206 13:56:03.603177 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:03 crc kubenswrapper[4706]: I1206 13:56:03.603193 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:03 crc kubenswrapper[4706]: I1206 13:56:03.603227 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:03 crc kubenswrapper[4706]: I1206 13:56:03.603239 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:03 crc kubenswrapper[4706]: I1206 13:56:03.603259 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:03 crc kubenswrapper[4706]: I1206 13:56:03.603246 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:03 crc kubenswrapper[4706]: I1206 13:56:03.694669 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 06 13:56:03 crc kubenswrapper[4706]: I1206 13:56:03.823931 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.390857 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.391038 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.392700 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.392764 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.392783 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.399810 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.604283 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.604496 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.604542 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.606106 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.606154 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.606171 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.606168 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.606296 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.606338 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.606170 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.606458 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:04 crc kubenswrapper[4706]: I1206 13:56:04.606485 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:05 crc kubenswrapper[4706]: I1206 13:56:05.467065 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 13:56:05 crc kubenswrapper[4706]: I1206 13:56:05.606762 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:05 crc kubenswrapper[4706]: I1206 13:56:05.608075 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:05 crc kubenswrapper[4706]: I1206 13:56:05.608155 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:05 crc kubenswrapper[4706]: I1206 13:56:05.608175 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:05 crc kubenswrapper[4706]: I1206 13:56:05.636907 4706 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 13:56:05 crc kubenswrapper[4706]: I1206 13:56:05.636983 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 13:56:07 crc kubenswrapper[4706]: E1206 13:56:07.600559 4706 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 13:56:08 crc kubenswrapper[4706]: I1206 13:56:08.764437 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 13:56:08 crc kubenswrapper[4706]: I1206 13:56:08.764688 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:08 crc kubenswrapper[4706]: I1206 13:56:08.766199 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:08 crc kubenswrapper[4706]: I1206 13:56:08.766347 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:08 crc kubenswrapper[4706]: I1206 13:56:08.766375 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:09 crc kubenswrapper[4706]: E1206 13:56:09.105790 4706 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 06 13:56:09 crc kubenswrapper[4706]: I1206 13:56:09.440027 4706 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 06 13:56:09 crc kubenswrapper[4706]: I1206 13:56:09.478825 4706 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 06 13:56:09 crc kubenswrapper[4706]: I1206 13:56:09.478882 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 06 13:56:10 crc kubenswrapper[4706]: E1206 13:56:10.447501 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 06 13:56:10 crc kubenswrapper[4706]: I1206 13:56:10.561423 4706 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Dec 06 13:56:10 crc kubenswrapper[4706]: I1206 13:56:10.561490 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 13:56:10 crc kubenswrapper[4706]: I1206 13:56:10.566394 4706 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Dec 06 13:56:10 crc kubenswrapper[4706]: I1206 13:56:10.566430 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 13:56:10 crc kubenswrapper[4706]: I1206 13:56:10.706134 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:10 crc kubenswrapper[4706]: I1206 13:56:10.707761 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:10 crc kubenswrapper[4706]: I1206 13:56:10.707901 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:10 crc kubenswrapper[4706]: I1206 13:56:10.707995 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:10 crc kubenswrapper[4706]: I1206 13:56:10.708096 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 13:56:10 crc kubenswrapper[4706]: I1206 13:56:10.913221 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 06 13:56:10 crc kubenswrapper[4706]: I1206 13:56:10.913565 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:10 crc kubenswrapper[4706]: I1206 13:56:10.914923 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:10 crc kubenswrapper[4706]: I1206 13:56:10.915095 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:10 crc kubenswrapper[4706]: I1206 13:56:10.915185 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:11 crc kubenswrapper[4706]: I1206 13:56:11.951841 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 13:56:11 crc kubenswrapper[4706]: I1206 13:56:11.952632 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:11 crc kubenswrapper[4706]: I1206 13:56:11.954047 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:11 crc kubenswrapper[4706]: I1206 13:56:11.954095 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:11 crc kubenswrapper[4706]: I1206 13:56:11.954113 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:13 crc kubenswrapper[4706]: I1206 13:56:13.835018 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:56:13 crc kubenswrapper[4706]: I1206 13:56:13.835357 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:13 crc kubenswrapper[4706]: I1206 13:56:13.837215 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:13 crc kubenswrapper[4706]: I1206 13:56:13.837276 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:13 crc kubenswrapper[4706]: I1206 13:56:13.837302 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:13 crc kubenswrapper[4706]: I1206 13:56:13.843613 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:56:14 crc kubenswrapper[4706]: I1206 13:56:14.631780 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 13:56:14 crc kubenswrapper[4706]: I1206 13:56:14.633154 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:14 crc kubenswrapper[4706]: I1206 13:56:14.633207 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:14 crc kubenswrapper[4706]: I1206 13:56:14.633224 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.550774 4706 trace.go:236] Trace[2119054043]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 13:56:01.485) (total time: 14065ms): Dec 06 13:56:15 crc kubenswrapper[4706]: Trace[2119054043]: ---"Objects listed" error: 14065ms (13:56:15.550) Dec 06 13:56:15 crc kubenswrapper[4706]: Trace[2119054043]: [14.065345034s] [14.065345034s] END Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.550814 4706 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.551075 4706 trace.go:236] Trace[1641949969]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 13:56:00.837) (total time: 14713ms): Dec 06 13:56:15 crc kubenswrapper[4706]: Trace[1641949969]: ---"Objects listed" error: 14713ms (13:56:15.550) Dec 06 13:56:15 crc kubenswrapper[4706]: Trace[1641949969]: [14.713323765s] [14.713323765s] END Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.551111 4706 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.553558 4706 trace.go:236] Trace[1637682981]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 13:56:01.191) (total time: 14361ms): Dec 06 13:56:15 crc kubenswrapper[4706]: Trace[1637682981]: ---"Objects listed" error: 14361ms (13:56:15.553) Dec 06 13:56:15 crc kubenswrapper[4706]: Trace[1637682981]: [14.361784998s] [14.361784998s] END Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.553597 4706 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.555980 4706 trace.go:236] Trace[488424917]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 13:56:01.758) (total time: 13797ms): Dec 06 13:56:15 crc kubenswrapper[4706]: Trace[488424917]: ---"Objects listed" error: 13796ms (13:56:15.555) Dec 06 13:56:15 crc kubenswrapper[4706]: Trace[488424917]: [13.797028218s] [13.797028218s] END Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.556046 4706 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.557493 4706 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.634350 4706 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:53802->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.634415 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:53802->192.168.126.11:17697: read: connection reset by peer" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.635277 4706 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.635350 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.637387 4706 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.637449 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.718650 4706 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.718740 4706 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.719747 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.719776 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.719785 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.719800 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.719810 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:15Z","lastTransitionTime":"2025-12-06T13:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:15 crc kubenswrapper[4706]: E1206 13:56:15.734832 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.738895 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.738933 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.738946 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.738962 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.738973 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:15Z","lastTransitionTime":"2025-12-06T13:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:15 crc kubenswrapper[4706]: E1206 13:56:15.749459 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.753334 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.753389 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.753410 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.753431 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.753448 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:15Z","lastTransitionTime":"2025-12-06T13:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:15 crc kubenswrapper[4706]: E1206 13:56:15.772899 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.777965 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.778018 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.778037 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.778060 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.778078 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:15Z","lastTransitionTime":"2025-12-06T13:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:15 crc kubenswrapper[4706]: E1206 13:56:15.792869 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.797148 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.797200 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.797220 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.797244 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.797263 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:15Z","lastTransitionTime":"2025-12-06T13:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:15 crc kubenswrapper[4706]: E1206 13:56:15.814783 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:15 crc kubenswrapper[4706]: E1206 13:56:15.814898 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.817167 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.817225 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.817243 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.817277 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.817294 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:15Z","lastTransitionTime":"2025-12-06T13:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.919325 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.919401 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.919420 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.919438 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:15 crc kubenswrapper[4706]: I1206 13:56:15.919452 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:15Z","lastTransitionTime":"2025-12-06T13:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.022475 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.022515 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.022528 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.022545 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.022557 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:16Z","lastTransitionTime":"2025-12-06T13:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.124380 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.124412 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.124439 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.124453 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.124461 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:16Z","lastTransitionTime":"2025-12-06T13:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.226923 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.226965 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.226977 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.226994 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.227006 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:16Z","lastTransitionTime":"2025-12-06T13:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.329228 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.329269 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.329280 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.329298 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.329324 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:16Z","lastTransitionTime":"2025-12-06T13:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.431198 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.431233 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.431242 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.431256 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.431264 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:16Z","lastTransitionTime":"2025-12-06T13:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.441733 4706 apiserver.go:52] "Watching apiserver" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.445540 4706 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.445802 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.446161 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.446197 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 13:56:16 crc kubenswrapper[4706]: E1206 13:56:16.446221 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.446283 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.446321 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.446387 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.447276 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:16 crc kubenswrapper[4706]: E1206 13:56:16.447432 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:16 crc kubenswrapper[4706]: E1206 13:56:16.447484 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.448162 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.448363 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.448650 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.448732 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.448857 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.448895 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.450061 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.452940 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.452944 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.465202 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.477564 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.487712 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.499973 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.510163 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.520408 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.531339 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.533055 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.533082 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.533091 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.533103 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.533128 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:16Z","lastTransitionTime":"2025-12-06T13:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.542482 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.543054 4706 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.562813 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.562853 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.562873 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.562895 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.562913 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.562927 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.562945 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.562968 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.562989 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563012 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563028 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563043 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563094 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563112 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563129 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563149 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563187 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563206 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563222 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563226 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563237 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563240 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563254 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563293 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563346 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563367 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563384 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563401 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563416 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563432 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563448 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563466 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563483 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563555 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563568 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563572 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563637 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563656 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563671 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563688 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563706 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563721 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563736 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563748 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563751 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563781 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563797 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563813 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563827 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563841 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563831 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563856 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563890 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563921 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563944 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563964 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563982 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.563998 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564014 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564036 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564060 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564078 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564096 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564112 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564133 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564150 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564169 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564188 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564205 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564221 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564238 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564272 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564294 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564327 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564346 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564363 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564378 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564394 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564412 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564427 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564447 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564462 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564480 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564498 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564514 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564533 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564548 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564562 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564580 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564595 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564613 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564630 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564645 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564662 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564677 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564694 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564712 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564727 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564747 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564763 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564779 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564799 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564815 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564831 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564848 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564864 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564918 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564937 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564954 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564971 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564989 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565005 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565026 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565050 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565067 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565083 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565101 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565118 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565134 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565150 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565168 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565183 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565199 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565218 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565234 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565249 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565266 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565284 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565302 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565333 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565350 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565367 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565394 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564057 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564056 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564254 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564302 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564342 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564354 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564459 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564471 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564486 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564605 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564605 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564611 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564644 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564724 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564783 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564841 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564855 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564937 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.564987 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565129 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565178 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565246 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565326 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565397 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565408 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565565 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565675 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565691 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565750 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565744 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565414 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565792 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565816 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565828 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565852 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565876 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565893 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565910 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565926 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565929 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565943 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565945 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565945 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.565962 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566034 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566055 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566079 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566108 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566135 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566157 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566173 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566190 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566208 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566227 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566242 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566258 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566276 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566294 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566361 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566386 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566406 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566429 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566453 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566479 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566501 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566522 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566545 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566571 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566604 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566633 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566658 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566681 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566708 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566732 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566756 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566778 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566797 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566813 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566829 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566845 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566861 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566878 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566898 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566915 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566936 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566957 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566973 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566993 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567017 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567036 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567053 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567069 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567089 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567106 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567129 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567148 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567164 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567180 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567196 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567213 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567229 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567257 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567276 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567292 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567347 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567368 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567402 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567420 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567440 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567461 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567483 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567501 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567521 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567538 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567553 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567570 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567595 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567614 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567669 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567681 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567691 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567701 4706 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567711 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567720 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567730 4706 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567740 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567750 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567759 4706 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567769 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567779 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567788 4706 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566108 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566115 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566151 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566166 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566214 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566241 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566251 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566376 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566380 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566395 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566415 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566441 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566501 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566861 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.566994 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.568521 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567622 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567648 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: E1206 13:56:16.567812 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:56:17.067792933 +0000 UTC m=+19.903585790 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.568782 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567831 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.567881 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.568069 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.568090 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.568127 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.568174 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.568499 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.568560 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.568863 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.568709 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569136 4706 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569164 4706 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569184 4706 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569203 4706 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569249 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569282 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569299 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569345 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569362 4706 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569380 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569398 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569417 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569434 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569451 4706 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569469 4706 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569485 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569502 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569521 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569538 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569557 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569575 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569591 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569610 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569627 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569643 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569659 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569675 4706 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569692 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569708 4706 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.569841 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.570627 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.570705 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.570765 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.571228 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.571401 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.571448 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.571497 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.571956 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.571973 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 13:56:16 crc kubenswrapper[4706]: E1206 13:56:16.572098 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 13:56:16 crc kubenswrapper[4706]: E1206 13:56:16.572404 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:17.072389033 +0000 UTC m=+19.908181890 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.572401 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: E1206 13:56:16.572486 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 13:56:16 crc kubenswrapper[4706]: E1206 13:56:16.572545 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:17.072531137 +0000 UTC m=+19.908323994 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.572226 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.572243 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.572244 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.572253 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.572105 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.572166 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.572614 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.572621 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.572934 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.573035 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.573052 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.573269 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.573377 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.573413 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.573454 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.573478 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.573460 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.573728 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.573889 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.573982 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.574150 4706 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.574159 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.574125 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.574495 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.574710 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.574809 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.574961 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.574985 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.575091 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.575338 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.575454 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.575625 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.575663 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.575676 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.575854 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.575939 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.576105 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.576202 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.576244 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.576576 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.579438 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.591182 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.591758 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.591821 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.592884 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 13:56:16 crc kubenswrapper[4706]: E1206 13:56:16.594952 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 13:56:16 crc kubenswrapper[4706]: E1206 13:56:16.594999 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 13:56:16 crc kubenswrapper[4706]: E1206 13:56:16.595013 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:16 crc kubenswrapper[4706]: E1206 13:56:16.595115 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:17.095087958 +0000 UTC m=+19.930880925 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:16 crc kubenswrapper[4706]: E1206 13:56:16.595196 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 13:56:16 crc kubenswrapper[4706]: E1206 13:56:16.595208 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 13:56:16 crc kubenswrapper[4706]: E1206 13:56:16.595218 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:16 crc kubenswrapper[4706]: E1206 13:56:16.595277 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:17.095266643 +0000 UTC m=+19.931059520 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.597029 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.597654 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.597803 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.597942 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.599137 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.599563 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.599818 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.600039 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.600566 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.600666 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.602477 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.602712 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.602783 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.603114 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.603154 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.603447 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.604443 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.607562 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.608300 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.608489 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.608548 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.608584 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.608722 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.608735 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.608745 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.608894 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.609118 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.609189 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.609204 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.609710 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.610141 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.610707 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.608846 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.611086 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.616737 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.616825 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.616887 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.617688 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.617755 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.619530 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.619842 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.619821 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.619955 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.620035 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.620175 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.620457 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.620481 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.620650 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.620745 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.620957 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.621091 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.621598 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.623604 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.623594 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.623647 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.623977 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.624123 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.624164 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.624170 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.624292 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.624433 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.624478 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.624568 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.624910 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.625133 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.625258 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.625440 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.625446 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.625708 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.625874 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.626768 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.628242 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.630785 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.637160 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.637194 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.637211 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.637227 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.637237 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:16Z","lastTransitionTime":"2025-12-06T13:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.638570 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.641621 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3" exitCode=255 Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.641655 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3"} Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.642527 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.647926 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.656206 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.670879 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.670943 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671031 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671051 4706 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671071 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671086 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671102 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671119 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671131 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671251 4706 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671265 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671277 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671288 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671299 4706 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671332 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671343 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671354 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671366 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671379 4706 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671391 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671404 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671416 4706 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671430 4706 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671442 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671454 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671467 4706 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671479 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671491 4706 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671503 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671516 4706 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671528 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671540 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671553 4706 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671565 4706 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671577 4706 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671589 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671601 4706 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671613 4706 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671626 4706 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671637 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671647 4706 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671659 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671671 4706 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671683 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671771 4706 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671785 4706 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671797 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671808 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671820 4706 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671832 4706 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671844 4706 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671855 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671869 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671881 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671893 4706 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671905 4706 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671916 4706 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671932 4706 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671947 4706 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671960 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671971 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671983 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.671994 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672013 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672024 4706 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672035 4706 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672047 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672059 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672070 4706 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672112 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672126 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672137 4706 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672148 4706 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672159 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672170 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672183 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672194 4706 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672204 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672216 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672228 4706 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672240 4706 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672252 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672264 4706 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672276 4706 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672287 4706 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672298 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672330 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672342 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672354 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672365 4706 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672376 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672387 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672398 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672408 4706 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672419 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672430 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672441 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672452 4706 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672464 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672476 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672486 4706 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672499 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672512 4706 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672523 4706 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672535 4706 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672546 4706 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672557 4706 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672569 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672580 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672592 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672603 4706 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672614 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672625 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672637 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672648 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672659 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672671 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672682 4706 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672693 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672704 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672716 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672727 4706 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672738 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672750 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672761 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672772 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672783 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672793 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672804 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672814 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672826 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672836 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672847 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672858 4706 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672870 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672881 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672893 4706 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672904 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672917 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672928 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672939 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672950 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672961 4706 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672973 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672985 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.672996 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.673048 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.673067 4706 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.673081 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.673094 4706 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.673106 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.673117 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.673129 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.673111 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.673141 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.673300 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.673355 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.699152 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.709765 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.709975 4706 scope.go:117] "RemoveContainer" containerID="09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.732465 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.742937 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.742981 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.743006 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.743022 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.743031 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:16Z","lastTransitionTime":"2025-12-06T13:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.747516 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.759155 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.761386 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.770548 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.770591 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.772451 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 13:56:16 crc kubenswrapper[4706]: W1206 13:56:16.804057 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-e4f52f237d34c45b95c963f4271e1450ff64d177ab68aa41cb9332369f500e64 WatchSource:0}: Error finding container e4f52f237d34c45b95c963f4271e1450ff64d177ab68aa41cb9332369f500e64: Status 404 returned error can't find the container with id e4f52f237d34c45b95c963f4271e1450ff64d177ab68aa41cb9332369f500e64 Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.845362 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.845394 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.845402 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.845414 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.845424 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:16Z","lastTransitionTime":"2025-12-06T13:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.947639 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.947681 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.947690 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.947706 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:16 crc kubenswrapper[4706]: I1206 13:56:16.947717 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:16Z","lastTransitionTime":"2025-12-06T13:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.050417 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.050454 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.050464 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.050479 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.050513 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:17Z","lastTransitionTime":"2025-12-06T13:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.075894 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.075980 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.076029 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:17 crc kubenswrapper[4706]: E1206 13:56:17.076088 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:56:18.07606045 +0000 UTC m=+20.911853307 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:56:17 crc kubenswrapper[4706]: E1206 13:56:17.076130 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 13:56:17 crc kubenswrapper[4706]: E1206 13:56:17.076133 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 13:56:17 crc kubenswrapper[4706]: E1206 13:56:17.076195 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:18.076175953 +0000 UTC m=+20.911968860 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 13:56:17 crc kubenswrapper[4706]: E1206 13:56:17.076223 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:18.076203114 +0000 UTC m=+20.911995971 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.130693 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-67v56"] Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.131021 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-49gn6"] Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.131135 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-9rrsc"] Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.131165 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.131278 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-49gn6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.131900 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.133687 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.134281 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.135148 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.135352 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.135467 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.135578 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.135686 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.135859 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.135966 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.138976 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.139961 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.140052 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.140096 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.146285 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.153257 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.153280 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.153289 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.153301 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.153327 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:17Z","lastTransitionTime":"2025-12-06T13:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.156073 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.164101 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.177169 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsw72\" (UniqueName: \"kubernetes.io/projected/904f5a16-0c23-4a2a-a945-59b1142cb4e1-kube-api-access-dsw72\") pod \"machine-config-daemon-67v56\" (UID: \"904f5a16-0c23-4a2a-a945-59b1142cb4e1\") " pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.177226 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-os-release\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.177249 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4fab411c-d62e-4003-b184-77b45c252e9f-hosts-file\") pod \"node-resolver-49gn6\" (UID: \"4fab411c-d62e-4003-b184-77b45c252e9f\") " pod="openshift-dns/node-resolver-49gn6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.177284 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.177320 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/904f5a16-0c23-4a2a-a945-59b1142cb4e1-proxy-tls\") pod \"machine-config-daemon-67v56\" (UID: \"904f5a16-0c23-4a2a-a945-59b1142cb4e1\") " pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.177341 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhb9z\" (UniqueName: \"kubernetes.io/projected/4fab411c-d62e-4003-b184-77b45c252e9f-kube-api-access-fhb9z\") pod \"node-resolver-49gn6\" (UID: \"4fab411c-d62e-4003-b184-77b45c252e9f\") " pod="openshift-dns/node-resolver-49gn6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.177367 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.177399 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/904f5a16-0c23-4a2a-a945-59b1142cb4e1-rootfs\") pod \"machine-config-daemon-67v56\" (UID: \"904f5a16-0c23-4a2a-a945-59b1142cb4e1\") " pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.177419 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-system-cni-dir\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: E1206 13:56:17.177434 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 13:56:17 crc kubenswrapper[4706]: E1206 13:56:17.177464 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 13:56:17 crc kubenswrapper[4706]: E1206 13:56:17.177477 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:17 crc kubenswrapper[4706]: E1206 13:56:17.177492 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 13:56:17 crc kubenswrapper[4706]: E1206 13:56:17.177524 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.177442 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-cni-binary-copy\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: E1206 13:56:17.177540 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:17 crc kubenswrapper[4706]: E1206 13:56:17.177525 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:18.17750778 +0000 UTC m=+21.013300637 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.177622 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7dm8\" (UniqueName: \"kubernetes.io/projected/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-kube-api-access-d7dm8\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: E1206 13:56:17.177675 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:18.177626873 +0000 UTC m=+21.013419730 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.177764 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/904f5a16-0c23-4a2a-a945-59b1142cb4e1-mcd-auth-proxy-config\") pod \"machine-config-daemon-67v56\" (UID: \"904f5a16-0c23-4a2a-a945-59b1142cb4e1\") " pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.177788 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.177813 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-cnibin\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.177833 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.181143 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.193464 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.200997 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.215819 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.234243 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.244871 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.252904 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.255493 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.255536 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.255546 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.255565 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.255574 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:17Z","lastTransitionTime":"2025-12-06T13:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.262981 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.271612 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.278471 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/904f5a16-0c23-4a2a-a945-59b1142cb4e1-mcd-auth-proxy-config\") pod \"machine-config-daemon-67v56\" (UID: \"904f5a16-0c23-4a2a-a945-59b1142cb4e1\") " pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.278525 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.278546 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-cnibin\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.278568 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.278589 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsw72\" (UniqueName: \"kubernetes.io/projected/904f5a16-0c23-4a2a-a945-59b1142cb4e1-kube-api-access-dsw72\") pod \"machine-config-daemon-67v56\" (UID: \"904f5a16-0c23-4a2a-a945-59b1142cb4e1\") " pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.278610 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-os-release\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.278627 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4fab411c-d62e-4003-b184-77b45c252e9f-hosts-file\") pod \"node-resolver-49gn6\" (UID: \"4fab411c-d62e-4003-b184-77b45c252e9f\") " pod="openshift-dns/node-resolver-49gn6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.278662 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/904f5a16-0c23-4a2a-a945-59b1142cb4e1-proxy-tls\") pod \"machine-config-daemon-67v56\" (UID: \"904f5a16-0c23-4a2a-a945-59b1142cb4e1\") " pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.278660 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-cnibin\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.278706 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhb9z\" (UniqueName: \"kubernetes.io/projected/4fab411c-d62e-4003-b184-77b45c252e9f-kube-api-access-fhb9z\") pod \"node-resolver-49gn6\" (UID: \"4fab411c-d62e-4003-b184-77b45c252e9f\") " pod="openshift-dns/node-resolver-49gn6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.278784 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/904f5a16-0c23-4a2a-a945-59b1142cb4e1-rootfs\") pod \"machine-config-daemon-67v56\" (UID: \"904f5a16-0c23-4a2a-a945-59b1142cb4e1\") " pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.278803 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-system-cni-dir\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.278832 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-cni-binary-copy\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.278851 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7dm8\" (UniqueName: \"kubernetes.io/projected/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-kube-api-access-d7dm8\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.278895 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.278956 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-system-cni-dir\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.278989 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/904f5a16-0c23-4a2a-a945-59b1142cb4e1-rootfs\") pod \"machine-config-daemon-67v56\" (UID: \"904f5a16-0c23-4a2a-a945-59b1142cb4e1\") " pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.279028 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4fab411c-d62e-4003-b184-77b45c252e9f-hosts-file\") pod \"node-resolver-49gn6\" (UID: \"4fab411c-d62e-4003-b184-77b45c252e9f\") " pod="openshift-dns/node-resolver-49gn6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.279078 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-os-release\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.279267 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/904f5a16-0c23-4a2a-a945-59b1142cb4e1-mcd-auth-proxy-config\") pod \"machine-config-daemon-67v56\" (UID: \"904f5a16-0c23-4a2a-a945-59b1142cb4e1\") " pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.279478 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.279576 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-cni-binary-copy\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.282917 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/904f5a16-0c23-4a2a-a945-59b1142cb4e1-proxy-tls\") pod \"machine-config-daemon-67v56\" (UID: \"904f5a16-0c23-4a2a-a945-59b1142cb4e1\") " pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.285591 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.296670 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsw72\" (UniqueName: \"kubernetes.io/projected/904f5a16-0c23-4a2a-a945-59b1142cb4e1-kube-api-access-dsw72\") pod \"machine-config-daemon-67v56\" (UID: \"904f5a16-0c23-4a2a-a945-59b1142cb4e1\") " pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.297605 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.297839 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhb9z\" (UniqueName: \"kubernetes.io/projected/4fab411c-d62e-4003-b184-77b45c252e9f-kube-api-access-fhb9z\") pod \"node-resolver-49gn6\" (UID: \"4fab411c-d62e-4003-b184-77b45c252e9f\") " pod="openshift-dns/node-resolver-49gn6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.300532 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7dm8\" (UniqueName: \"kubernetes.io/projected/80e25f9c-2df9-42c2-b82a-ecc877b7fd32-kube-api-access-d7dm8\") pod \"multus-additional-cni-plugins-9rrsc\" (UID: \"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\") " pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.309472 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.322038 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.339578 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.357893 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.357945 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.357954 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.357970 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.357983 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:17Z","lastTransitionTime":"2025-12-06T13:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.358481 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.447492 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.454020 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-49gn6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.463368 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.475514 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.475575 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.475588 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.475607 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.475626 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:17Z","lastTransitionTime":"2025-12-06T13:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.510937 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.510941 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:17 crc kubenswrapper[4706]: E1206 13:56:17.511152 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:17 crc kubenswrapper[4706]: E1206 13:56:17.511326 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.547181 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.548076 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.567230 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.577238 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.578387 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.578413 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.578424 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.578439 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.578449 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:17Z","lastTransitionTime":"2025-12-06T13:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.589972 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.601725 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.602422 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.603641 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.604266 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.605248 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.605943 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.617484 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.618342 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.619135 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.622065 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.623510 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.627202 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.627727 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.630351 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.631278 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.631844 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.635170 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.635785 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.636454 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.636613 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.637914 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.639796 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.649556 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.650508 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.650660 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.651548 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.653035 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.654127 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.655303 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.656525 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.657724 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.658608 4706 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.658729 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.658749 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.661391 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.662723 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.663502 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.665621 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.666686 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.667964 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.670134 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.671847 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.672739 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.672985 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.674029 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.674955 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.676465 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.677091 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.678359 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.679024 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.680495 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.681077 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.683132 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.684205 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.684620 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.684652 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.684662 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.684678 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.684690 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:17Z","lastTransitionTime":"2025-12-06T13:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.685853 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.688198 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.690701 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.691362 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-q6bgc"] Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.700720 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.700777 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerStarted","Data":"4ee36580838003d77f3253f361766bd285128c1bb264aaeb423ae5d482757fd0"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.700802 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.700816 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-xqlx6"] Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.701104 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.701165 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.702340 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-49gn6" event={"ID":"4fab411c-d62e-4003-b184-77b45c252e9f","Type":"ContainerStarted","Data":"2fb7024dea8fad454a8a6a3537d9e8937756bcce1c5f91c990f6048c069cf67e"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.702396 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.702430 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.702448 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e4f52f237d34c45b95c963f4271e1450ff64d177ab68aa41cb9332369f500e64"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.702464 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.702484 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b484698a0c840c07f476811e27a6570a5cf2065803eb15ecb92088b614dbad85"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.702506 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"f7a77f252ca7dcc4875f89942c1cc2e72dc431a8a62ce19e46d33e04c6a80da2"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.702525 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" event={"ID":"80e25f9c-2df9-42c2-b82a-ecc877b7fd32","Type":"ContainerStarted","Data":"43dba13dc86518e48ec7b5c97142fc4cfd2274b170823ffc81bef59a0208f899"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.702844 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.706302 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.706744 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.707000 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.707115 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.707410 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.707564 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.708351 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.708477 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.708527 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.720725 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.738737 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.751628 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.767869 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.782635 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-systemd-units\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.782693 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovnkube-config\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.782718 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovnkube-script-lib\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.782746 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-run-k8s-cni-cncf-io\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.782794 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-multus-cni-dir\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.782812 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-env-overrides\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.782832 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-run-netns\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.782848 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-ovn\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.782870 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-multus-socket-dir-parent\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.782902 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-os-release\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.782922 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thxfp\" (UniqueName: \"kubernetes.io/projected/d3fdd6dc-f817-486c-b0ee-18a6f2185559-kube-api-access-thxfp\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.782943 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-cni-netd\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.782984 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnr6h\" (UniqueName: \"kubernetes.io/projected/c6db0d03-fbc0-4b3b-b89d-282767c274be-kube-api-access-fnr6h\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783001 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d3fdd6dc-f817-486c-b0ee-18a6f2185559-cni-binary-copy\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783020 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-cni-bin\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783045 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-kubelet\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783064 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-run-netns\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783089 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-node-log\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783114 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-run-ovn-kubernetes\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783137 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783159 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-cnibin\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783179 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-var-lib-kubelet\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783197 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-etc-kubernetes\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783214 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-etc-openvswitch\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783231 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-system-cni-dir\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783255 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-slash\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783272 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d3fdd6dc-f817-486c-b0ee-18a6f2185559-multus-daemon-config\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783287 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-run-multus-certs\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783322 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-var-lib-openvswitch\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783338 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-openvswitch\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783353 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-systemd\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783370 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-log-socket\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783391 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-var-lib-cni-bin\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783406 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-multus-conf-dir\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783452 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-var-lib-cni-multus\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783472 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-hostroot\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.783495 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovn-node-metrics-cert\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.784161 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.786943 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.786992 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.787003 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.787019 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.787032 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:17Z","lastTransitionTime":"2025-12-06T13:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.800976 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.817602 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.833093 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.847598 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.865102 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.879442 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884203 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-systemd-units\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884248 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovnkube-config\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884268 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovnkube-script-lib\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884296 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-run-k8s-cni-cncf-io\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884330 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-multus-cni-dir\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884345 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-env-overrides\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884361 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-systemd-units\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884451 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-multus-cni-dir\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884442 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-run-k8s-cni-cncf-io\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884368 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-run-netns\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884595 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-ovn\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884619 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-multus-socket-dir-parent\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884652 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-ovn\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884696 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-os-release\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884712 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thxfp\" (UniqueName: \"kubernetes.io/projected/d3fdd6dc-f817-486c-b0ee-18a6f2185559-kube-api-access-thxfp\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884728 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-cni-netd\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884743 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnr6h\" (UniqueName: \"kubernetes.io/projected/c6db0d03-fbc0-4b3b-b89d-282767c274be-kube-api-access-fnr6h\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884756 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d3fdd6dc-f817-486c-b0ee-18a6f2185559-cni-binary-copy\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884775 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-cni-bin\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884800 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-kubelet\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884816 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-run-netns\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884831 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-node-log\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884847 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-run-ovn-kubernetes\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884864 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884885 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-cnibin\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884901 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-var-lib-kubelet\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884918 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-etc-kubernetes\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884940 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-etc-openvswitch\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884956 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-system-cni-dir\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884971 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-slash\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.884988 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d3fdd6dc-f817-486c-b0ee-18a6f2185559-multus-daemon-config\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885002 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-run-multus-certs\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885018 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-var-lib-openvswitch\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885033 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-openvswitch\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885048 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-systemd\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885063 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-log-socket\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885081 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-var-lib-cni-bin\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885096 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-multus-conf-dir\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885112 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-var-lib-cni-multus\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885127 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-hostroot\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885142 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovnkube-config\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885143 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovn-node-metrics-cert\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885229 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovnkube-script-lib\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885261 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-var-lib-kubelet\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885296 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-var-lib-openvswitch\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885330 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-multus-socket-dir-parent\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885353 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-run-netns\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885348 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-env-overrides\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885381 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-slash\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885416 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-os-release\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885443 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-run-netns\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885493 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-cnibin\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885519 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-etc-kubernetes\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885538 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-etc-openvswitch\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885571 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-system-cni-dir\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885608 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-cni-bin\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885633 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-cni-netd\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885778 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-kubelet\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885812 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-node-log\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885851 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-run-ovn-kubernetes\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885875 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d3fdd6dc-f817-486c-b0ee-18a6f2185559-multus-daemon-config\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885997 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-openvswitch\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.885998 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-var-lib-cni-bin\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.886018 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.886033 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-var-lib-cni-multus\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.886055 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-multus-conf-dir\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.886059 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-hostroot\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.886072 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-systemd\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.886086 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-log-socket\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.886145 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d3fdd6dc-f817-486c-b0ee-18a6f2185559-host-run-multus-certs\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.886350 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d3fdd6dc-f817-486c-b0ee-18a6f2185559-cni-binary-copy\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.890522 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovn-node-metrics-cert\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.893299 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.893405 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.893423 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.893448 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.893466 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:17Z","lastTransitionTime":"2025-12-06T13:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.903851 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.904873 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnr6h\" (UniqueName: \"kubernetes.io/projected/c6db0d03-fbc0-4b3b-b89d-282767c274be-kube-api-access-fnr6h\") pod \"ovnkube-node-q6bgc\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.905450 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thxfp\" (UniqueName: \"kubernetes.io/projected/d3fdd6dc-f817-486c-b0ee-18a6f2185559-kube-api-access-thxfp\") pod \"multus-xqlx6\" (UID: \"d3fdd6dc-f817-486c-b0ee-18a6f2185559\") " pod="openshift-multus/multus-xqlx6" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.939798 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.968247 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.996221 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.996265 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.996274 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.996288 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:17 crc kubenswrapper[4706]: I1206 13:56:17.996297 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:17Z","lastTransitionTime":"2025-12-06T13:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.030817 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:18 crc kubenswrapper[4706]: W1206 13:56:18.043349 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6db0d03_fbc0_4b3b_b89d_282767c274be.slice/crio-52e8d9434c7c56ec1bb8c5241162afeca03b248ddca4b5d72091792df4a71ec0 WatchSource:0}: Error finding container 52e8d9434c7c56ec1bb8c5241162afeca03b248ddca4b5d72091792df4a71ec0: Status 404 returned error can't find the container with id 52e8d9434c7c56ec1bb8c5241162afeca03b248ddca4b5d72091792df4a71ec0 Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.045063 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xqlx6" Dec 06 13:56:18 crc kubenswrapper[4706]: W1206 13:56:18.075668 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3fdd6dc_f817_486c_b0ee_18a6f2185559.slice/crio-2c95948d7187b0c8e6506da41512dcb2c055a3591c27c74d1b37302eb956cd00 WatchSource:0}: Error finding container 2c95948d7187b0c8e6506da41512dcb2c055a3591c27c74d1b37302eb956cd00: Status 404 returned error can't find the container with id 2c95948d7187b0c8e6506da41512dcb2c055a3591c27c74d1b37302eb956cd00 Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.086555 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.086683 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.086710 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:18 crc kubenswrapper[4706]: E1206 13:56:18.086807 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 13:56:18 crc kubenswrapper[4706]: E1206 13:56:18.086856 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:20.086841923 +0000 UTC m=+22.922634780 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 13:56:18 crc kubenswrapper[4706]: E1206 13:56:18.086968 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:56:20.086927955 +0000 UTC m=+22.922720812 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:56:18 crc kubenswrapper[4706]: E1206 13:56:18.087092 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 13:56:18 crc kubenswrapper[4706]: E1206 13:56:18.087162 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:20.087152901 +0000 UTC m=+22.922945988 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.104239 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.104269 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.104278 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.104296 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.104323 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:18Z","lastTransitionTime":"2025-12-06T13:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.187153 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.187224 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:18 crc kubenswrapper[4706]: E1206 13:56:18.187343 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 13:56:18 crc kubenswrapper[4706]: E1206 13:56:18.187359 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 13:56:18 crc kubenswrapper[4706]: E1206 13:56:18.187370 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:18 crc kubenswrapper[4706]: E1206 13:56:18.187409 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:20.187395649 +0000 UTC m=+23.023188506 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:18 crc kubenswrapper[4706]: E1206 13:56:18.187453 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 13:56:18 crc kubenswrapper[4706]: E1206 13:56:18.187461 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 13:56:18 crc kubenswrapper[4706]: E1206 13:56:18.187468 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:18 crc kubenswrapper[4706]: E1206 13:56:18.187486 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:20.187480262 +0000 UTC m=+23.023273109 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.208595 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.208634 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.208643 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.208657 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.208667 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:18Z","lastTransitionTime":"2025-12-06T13:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.313640 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.313994 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.314002 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.314016 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.314025 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:18Z","lastTransitionTime":"2025-12-06T13:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.416555 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.416620 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.416642 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.416666 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.416684 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:18Z","lastTransitionTime":"2025-12-06T13:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.510998 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:18 crc kubenswrapper[4706]: E1206 13:56:18.511134 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.519555 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.519586 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.519594 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.519608 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.519617 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:18Z","lastTransitionTime":"2025-12-06T13:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.622481 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.622528 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.622543 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.622562 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.622576 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:18Z","lastTransitionTime":"2025-12-06T13:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.694500 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-49gn6" event={"ID":"4fab411c-d62e-4003-b184-77b45c252e9f","Type":"ContainerStarted","Data":"a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d"} Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.695622 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerStarted","Data":"52e8d9434c7c56ec1bb8c5241162afeca03b248ddca4b5d72091792df4a71ec0"} Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.697181 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerStarted","Data":"8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf"} Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.697225 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerStarted","Data":"244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f"} Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.698762 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xqlx6" event={"ID":"d3fdd6dc-f817-486c-b0ee-18a6f2185559","Type":"ContainerStarted","Data":"2c95948d7187b0c8e6506da41512dcb2c055a3591c27c74d1b37302eb956cd00"} Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.700066 4706 generic.go:334] "Generic (PLEG): container finished" podID="80e25f9c-2df9-42c2-b82a-ecc877b7fd32" containerID="bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74" exitCode=0 Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.700140 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" event={"ID":"80e25f9c-2df9-42c2-b82a-ecc877b7fd32","Type":"ContainerDied","Data":"bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74"} Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.719977 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.730460 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.730489 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.730499 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.730519 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.730529 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:18Z","lastTransitionTime":"2025-12-06T13:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.736034 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.753838 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.769772 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.785435 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.801370 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.820675 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.833956 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.833994 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.834003 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.834022 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.834035 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:18Z","lastTransitionTime":"2025-12-06T13:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.843649 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.858442 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.877563 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.894077 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.913011 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.928848 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.936951 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.936994 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.937004 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.937029 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.937041 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:18Z","lastTransitionTime":"2025-12-06T13:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.945418 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.962606 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.974958 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:18 crc kubenswrapper[4706]: I1206 13:56:18.989318 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.015223 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.030883 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.041514 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.041570 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.041580 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.041601 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.041612 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:19Z","lastTransitionTime":"2025-12-06T13:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.043929 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.059697 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.086742 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.123046 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.139321 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.144252 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.144331 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.144348 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.144371 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.144384 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:19Z","lastTransitionTime":"2025-12-06T13:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.247050 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.247096 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.247106 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.247120 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.247128 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:19Z","lastTransitionTime":"2025-12-06T13:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.349764 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.349802 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.349810 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.349826 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.349836 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:19Z","lastTransitionTime":"2025-12-06T13:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.452998 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.453041 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.453050 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.453064 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.453075 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:19Z","lastTransitionTime":"2025-12-06T13:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.511053 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.511069 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:19 crc kubenswrapper[4706]: E1206 13:56:19.511181 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:19 crc kubenswrapper[4706]: E1206 13:56:19.511376 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.556166 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.556219 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.556233 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.556252 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.556264 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:19Z","lastTransitionTime":"2025-12-06T13:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.659029 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.659084 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.659093 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.659123 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.659135 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:19Z","lastTransitionTime":"2025-12-06T13:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.705676 4706 generic.go:334] "Generic (PLEG): container finished" podID="80e25f9c-2df9-42c2-b82a-ecc877b7fd32" containerID="eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6" exitCode=0 Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.705806 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" event={"ID":"80e25f9c-2df9-42c2-b82a-ecc877b7fd32","Type":"ContainerDied","Data":"eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6"} Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.707606 4706 generic.go:334] "Generic (PLEG): container finished" podID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerID="8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773" exitCode=0 Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.707693 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerDied","Data":"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773"} Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.709530 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353"} Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.711287 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xqlx6" event={"ID":"d3fdd6dc-f817-486c-b0ee-18a6f2185559","Type":"ContainerStarted","Data":"db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0"} Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.731954 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.751506 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.764029 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.764098 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.764111 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.764129 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.764170 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:19Z","lastTransitionTime":"2025-12-06T13:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.766147 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.794397 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.815133 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.835883 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.851866 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.866636 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.866679 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.866722 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.866743 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.866756 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:19Z","lastTransitionTime":"2025-12-06T13:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.869209 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.882531 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.900529 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.918709 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.937403 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.959377 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.969876 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.969922 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.969930 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.969944 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.969953 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:19Z","lastTransitionTime":"2025-12-06T13:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.975672 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.986765 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:19 crc kubenswrapper[4706]: I1206 13:56:19.998901 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:19Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.012544 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.035486 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.048791 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.099618 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.099655 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.099667 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.099684 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.099695 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:20Z","lastTransitionTime":"2025-12-06T13:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.107649 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.107805 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.107836 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:20 crc kubenswrapper[4706]: E1206 13:56:20.107965 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 13:56:20 crc kubenswrapper[4706]: E1206 13:56:20.108019 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:24.10800269 +0000 UTC m=+26.943795557 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 13:56:20 crc kubenswrapper[4706]: E1206 13:56:20.108075 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:56:24.108066991 +0000 UTC m=+26.943859858 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:56:20 crc kubenswrapper[4706]: E1206 13:56:20.108112 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 13:56:20 crc kubenswrapper[4706]: E1206 13:56:20.108137 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:24.108128583 +0000 UTC m=+26.943921450 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.111468 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.124612 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.142433 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.154980 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.180854 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.203826 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.203869 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.203885 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.203904 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.203916 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:20Z","lastTransitionTime":"2025-12-06T13:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.208670 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.208722 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:20 crc kubenswrapper[4706]: E1206 13:56:20.208827 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 13:56:20 crc kubenswrapper[4706]: E1206 13:56:20.208862 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 13:56:20 crc kubenswrapper[4706]: E1206 13:56:20.208874 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:20 crc kubenswrapper[4706]: E1206 13:56:20.208922 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:24.208905876 +0000 UTC m=+27.044698733 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:20 crc kubenswrapper[4706]: E1206 13:56:20.208835 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 13:56:20 crc kubenswrapper[4706]: E1206 13:56:20.208957 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 13:56:20 crc kubenswrapper[4706]: E1206 13:56:20.208968 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:20 crc kubenswrapper[4706]: E1206 13:56:20.209003 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:24.208989248 +0000 UTC m=+27.044782105 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.305638 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.306096 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.306107 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.306126 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.306138 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:20Z","lastTransitionTime":"2025-12-06T13:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.409286 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.409346 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.409358 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.409373 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.409384 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:20Z","lastTransitionTime":"2025-12-06T13:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.510422 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:20 crc kubenswrapper[4706]: E1206 13:56:20.510528 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.512174 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.512206 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.512215 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.512226 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.512235 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:20Z","lastTransitionTime":"2025-12-06T13:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.540108 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-2cc78"] Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.540519 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2cc78" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.542943 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.543459 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.544646 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.544939 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.558548 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.570542 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.590514 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.603079 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.611562 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8bz6\" (UniqueName: \"kubernetes.io/projected/aa340a3c-fe06-4af7-ab8d-77e223027038-kube-api-access-h8bz6\") pod \"node-ca-2cc78\" (UID: \"aa340a3c-fe06-4af7-ab8d-77e223027038\") " pod="openshift-image-registry/node-ca-2cc78" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.611634 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/aa340a3c-fe06-4af7-ab8d-77e223027038-serviceca\") pod \"node-ca-2cc78\" (UID: \"aa340a3c-fe06-4af7-ab8d-77e223027038\") " pod="openshift-image-registry/node-ca-2cc78" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.611663 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa340a3c-fe06-4af7-ab8d-77e223027038-host\") pod \"node-ca-2cc78\" (UID: \"aa340a3c-fe06-4af7-ab8d-77e223027038\") " pod="openshift-image-registry/node-ca-2cc78" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.615678 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.615734 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.615756 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.615785 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.615811 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:20Z","lastTransitionTime":"2025-12-06T13:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.616922 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.639495 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.652912 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.665698 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.679256 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.692118 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.709000 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.712539 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/aa340a3c-fe06-4af7-ab8d-77e223027038-serviceca\") pod \"node-ca-2cc78\" (UID: \"aa340a3c-fe06-4af7-ab8d-77e223027038\") " pod="openshift-image-registry/node-ca-2cc78" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.712619 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa340a3c-fe06-4af7-ab8d-77e223027038-host\") pod \"node-ca-2cc78\" (UID: \"aa340a3c-fe06-4af7-ab8d-77e223027038\") " pod="openshift-image-registry/node-ca-2cc78" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.712668 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8bz6\" (UniqueName: \"kubernetes.io/projected/aa340a3c-fe06-4af7-ab8d-77e223027038-kube-api-access-h8bz6\") pod \"node-ca-2cc78\" (UID: \"aa340a3c-fe06-4af7-ab8d-77e223027038\") " pod="openshift-image-registry/node-ca-2cc78" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.712766 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa340a3c-fe06-4af7-ab8d-77e223027038-host\") pod \"node-ca-2cc78\" (UID: \"aa340a3c-fe06-4af7-ab8d-77e223027038\") " pod="openshift-image-registry/node-ca-2cc78" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.713952 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/aa340a3c-fe06-4af7-ab8d-77e223027038-serviceca\") pod \"node-ca-2cc78\" (UID: \"aa340a3c-fe06-4af7-ab8d-77e223027038\") " pod="openshift-image-registry/node-ca-2cc78" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.717655 4706 generic.go:334] "Generic (PLEG): container finished" podID="80e25f9c-2df9-42c2-b82a-ecc877b7fd32" containerID="b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4" exitCode=0 Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.717709 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" event={"ID":"80e25f9c-2df9-42c2-b82a-ecc877b7fd32","Type":"ContainerDied","Data":"b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4"} Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.717869 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.717896 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.717908 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.717924 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.717935 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:20Z","lastTransitionTime":"2025-12-06T13:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.723971 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerStarted","Data":"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a"} Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.724028 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerStarted","Data":"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f"} Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.724050 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerStarted","Data":"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc"} Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.724068 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerStarted","Data":"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389"} Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.724086 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerStarted","Data":"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9"} Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.724106 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerStarted","Data":"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09"} Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.724640 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.740634 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.744265 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8bz6\" (UniqueName: \"kubernetes.io/projected/aa340a3c-fe06-4af7-ab8d-77e223027038-kube-api-access-h8bz6\") pod \"node-ca-2cc78\" (UID: \"aa340a3c-fe06-4af7-ab8d-77e223027038\") " pod="openshift-image-registry/node-ca-2cc78" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.751830 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.762785 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.774209 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.787750 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.801130 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.813571 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.820804 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.821718 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.822683 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.822763 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.822782 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:20Z","lastTransitionTime":"2025-12-06T13:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.831445 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.843566 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.855790 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2cc78" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.870977 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.890262 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.910672 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.925769 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.925952 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.926029 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.926134 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.926215 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:20Z","lastTransitionTime":"2025-12-06T13:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.932072 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.941006 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.962851 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.975834 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.976925 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.983916 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:20 crc kubenswrapper[4706]: I1206 13:56:20.993205 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:20Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.007221 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.024878 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.028592 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.028637 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.028652 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.028695 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.028711 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:21Z","lastTransitionTime":"2025-12-06T13:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.039719 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.051943 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.063387 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.082915 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.134195 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.134616 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.134638 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.134645 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.134659 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.134668 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:21Z","lastTransitionTime":"2025-12-06T13:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.154111 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.164899 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.177284 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.187260 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.196302 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.212597 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.223705 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.234002 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.236565 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.236599 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.236610 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.236624 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.236635 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:21Z","lastTransitionTime":"2025-12-06T13:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.271651 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.310900 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.338690 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.338928 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.339041 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.339174 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.339282 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:21Z","lastTransitionTime":"2025-12-06T13:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.357298 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.394971 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.432905 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.442032 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.442079 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.442091 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.442108 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.442121 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:21Z","lastTransitionTime":"2025-12-06T13:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.475667 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.510271 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:21 crc kubenswrapper[4706]: E1206 13:56:21.510436 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.510796 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:21 crc kubenswrapper[4706]: E1206 13:56:21.511079 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.515517 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.544898 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.544965 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.544976 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.544993 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.545008 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:21Z","lastTransitionTime":"2025-12-06T13:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.555379 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.595061 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.632577 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.647803 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.647841 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.647849 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.647862 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.647872 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:21Z","lastTransitionTime":"2025-12-06T13:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.730146 4706 generic.go:334] "Generic (PLEG): container finished" podID="80e25f9c-2df9-42c2-b82a-ecc877b7fd32" containerID="a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d" exitCode=0 Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.730194 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" event={"ID":"80e25f9c-2df9-42c2-b82a-ecc877b7fd32","Type":"ContainerDied","Data":"a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d"} Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.733407 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2cc78" event={"ID":"aa340a3c-fe06-4af7-ab8d-77e223027038","Type":"ContainerStarted","Data":"a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5"} Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.733431 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2cc78" event={"ID":"aa340a3c-fe06-4af7-ab8d-77e223027038","Type":"ContainerStarted","Data":"f9fc04f8ebcf90f36786c1db643aa78b6e0ae7af12f7eb339759e50d4da65fb1"} Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.752732 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.752777 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.752791 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.752812 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.752829 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:21Z","lastTransitionTime":"2025-12-06T13:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.763075 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.779242 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.790561 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.808862 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.838871 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.855702 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.855757 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.855769 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.855786 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.855797 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:21Z","lastTransitionTime":"2025-12-06T13:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.874000 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.912780 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.957994 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.958915 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.958963 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.958975 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.958992 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.959004 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:21Z","lastTransitionTime":"2025-12-06T13:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:21 crc kubenswrapper[4706]: I1206 13:56:21.998769 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:21Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.034770 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.062587 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.062822 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.062922 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.063057 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.063164 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:22Z","lastTransitionTime":"2025-12-06T13:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.083026 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.113711 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.154854 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.165087 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.165245 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.165407 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.165512 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.165589 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:22Z","lastTransitionTime":"2025-12-06T13:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.194141 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.232724 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.268429 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.268484 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.268501 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.268524 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.268543 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:22Z","lastTransitionTime":"2025-12-06T13:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.272147 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.332871 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.369083 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.370929 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.370955 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.370963 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.370977 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.370987 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:22Z","lastTransitionTime":"2025-12-06T13:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.395047 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.435085 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.473161 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.473464 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.473663 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.473816 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.473953 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:22Z","lastTransitionTime":"2025-12-06T13:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.476260 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.510539 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:22 crc kubenswrapper[4706]: E1206 13:56:22.510656 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.517912 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.554875 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.575707 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.575755 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.575771 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.575795 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.575813 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:22Z","lastTransitionTime":"2025-12-06T13:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.601106 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.640403 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.643494 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.649268 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.671184 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.678609 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.678839 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.678851 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.678893 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.678905 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:22Z","lastTransitionTime":"2025-12-06T13:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.706214 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.743241 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerStarted","Data":"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49"} Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.748426 4706 generic.go:334] "Generic (PLEG): container finished" podID="80e25f9c-2df9-42c2-b82a-ecc877b7fd32" containerID="4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5" exitCode=0 Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.748551 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" event={"ID":"80e25f9c-2df9-42c2-b82a-ecc877b7fd32","Type":"ContainerDied","Data":"4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5"} Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.762455 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.787533 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.787564 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.787571 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.787585 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.787597 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:22Z","lastTransitionTime":"2025-12-06T13:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.817136 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.833123 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.852906 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.890053 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.891004 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.891032 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.891044 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.891061 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.891070 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:22Z","lastTransitionTime":"2025-12-06T13:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.934877 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.970360 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:22Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.993216 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.993542 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.993552 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.993565 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:22 crc kubenswrapper[4706]: I1206 13:56:22.993574 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:22Z","lastTransitionTime":"2025-12-06T13:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.011729 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.058919 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.091738 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.095340 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.095365 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.095373 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.095387 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.095396 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:23Z","lastTransitionTime":"2025-12-06T13:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.140849 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.174910 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.198178 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.198206 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.198215 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.198227 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.198235 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:23Z","lastTransitionTime":"2025-12-06T13:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.214348 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.252818 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.300567 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.300601 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.300612 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.300654 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.300666 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:23Z","lastTransitionTime":"2025-12-06T13:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.302231 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.331806 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.371132 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.404241 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.404291 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.404352 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.404397 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.404416 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:23Z","lastTransitionTime":"2025-12-06T13:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.507415 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.507465 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.507480 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.507501 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.507516 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:23Z","lastTransitionTime":"2025-12-06T13:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.510946 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:23 crc kubenswrapper[4706]: E1206 13:56:23.511056 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.510954 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:23 crc kubenswrapper[4706]: E1206 13:56:23.511553 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.610600 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.610632 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.610640 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.610653 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.610662 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:23Z","lastTransitionTime":"2025-12-06T13:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.713719 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.713770 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.713786 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.713807 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.713824 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:23Z","lastTransitionTime":"2025-12-06T13:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.757248 4706 generic.go:334] "Generic (PLEG): container finished" podID="80e25f9c-2df9-42c2-b82a-ecc877b7fd32" containerID="c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16" exitCode=0 Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.757293 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" event={"ID":"80e25f9c-2df9-42c2-b82a-ecc877b7fd32","Type":"ContainerDied","Data":"c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16"} Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.795621 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.818163 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.818225 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.818244 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.818267 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.818285 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:23Z","lastTransitionTime":"2025-12-06T13:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.818909 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.857421 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.882550 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.897583 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.917884 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.925660 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.925701 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.925717 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.925739 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.925756 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:23Z","lastTransitionTime":"2025-12-06T13:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.930479 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.947937 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.960778 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.974876 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:23 crc kubenswrapper[4706]: I1206 13:56:23.991709 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:23Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.009825 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:24Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.030298 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.030376 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.030391 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.030411 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.030424 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:24Z","lastTransitionTime":"2025-12-06T13:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.034989 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:24Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.049659 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:24Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.067718 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:24Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.133632 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.133697 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.133714 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.133739 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.133758 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:24Z","lastTransitionTime":"2025-12-06T13:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.149699 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.149905 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:24 crc kubenswrapper[4706]: E1206 13:56:24.149947 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:56:32.149895062 +0000 UTC m=+34.985687949 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.150014 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:24 crc kubenswrapper[4706]: E1206 13:56:24.150029 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 13:56:24 crc kubenswrapper[4706]: E1206 13:56:24.150102 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:32.150079557 +0000 UTC m=+34.985872444 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 13:56:24 crc kubenswrapper[4706]: E1206 13:56:24.150202 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 13:56:24 crc kubenswrapper[4706]: E1206 13:56:24.150288 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:32.150267882 +0000 UTC m=+34.986060749 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.236895 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.236954 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.236971 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.236996 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.237015 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:24Z","lastTransitionTime":"2025-12-06T13:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.250681 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.250728 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:24 crc kubenswrapper[4706]: E1206 13:56:24.250841 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 13:56:24 crc kubenswrapper[4706]: E1206 13:56:24.250857 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 13:56:24 crc kubenswrapper[4706]: E1206 13:56:24.250868 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:24 crc kubenswrapper[4706]: E1206 13:56:24.250908 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:32.25089467 +0000 UTC m=+35.086687527 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:24 crc kubenswrapper[4706]: E1206 13:56:24.250971 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 13:56:24 crc kubenswrapper[4706]: E1206 13:56:24.251027 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 13:56:24 crc kubenswrapper[4706]: E1206 13:56:24.251054 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:24 crc kubenswrapper[4706]: E1206 13:56:24.251151 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:32.251121286 +0000 UTC m=+35.086914193 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.339699 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.339730 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.339739 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.339753 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.339761 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:24Z","lastTransitionTime":"2025-12-06T13:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.442064 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.442103 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.442113 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.442127 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.442136 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:24Z","lastTransitionTime":"2025-12-06T13:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.510106 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:24 crc kubenswrapper[4706]: E1206 13:56:24.510292 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.544911 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.544981 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.544998 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.545026 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.545048 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:24Z","lastTransitionTime":"2025-12-06T13:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.648947 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.649013 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.649030 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.649054 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.649078 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:24Z","lastTransitionTime":"2025-12-06T13:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.752255 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.752339 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.752358 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.752415 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.752437 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:24Z","lastTransitionTime":"2025-12-06T13:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.766206 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" event={"ID":"80e25f9c-2df9-42c2-b82a-ecc877b7fd32","Type":"ContainerStarted","Data":"95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb"} Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.786480 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:24Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.806378 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:24Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.840185 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:24Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.855769 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.855820 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.855842 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.855866 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.855884 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:24Z","lastTransitionTime":"2025-12-06T13:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.861802 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:24Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.875269 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:24Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.889863 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:24Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.902614 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:24Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.922401 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:24Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.939563 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:24Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.956282 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:24Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.958095 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.958152 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.958172 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.958197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.958213 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:24Z","lastTransitionTime":"2025-12-06T13:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.975413 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:24Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:24 crc kubenswrapper[4706]: I1206 13:56:24.993050 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:24Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.006827 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:25Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.025093 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:25Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.037105 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:25Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.060755 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.060810 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.060828 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.060857 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.060874 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:25Z","lastTransitionTime":"2025-12-06T13:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.164545 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.164625 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.164649 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.164679 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.164705 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:25Z","lastTransitionTime":"2025-12-06T13:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.267480 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.267529 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.267545 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.267567 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.267584 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:25Z","lastTransitionTime":"2025-12-06T13:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.369399 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.369446 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.369461 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.369481 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.369495 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:25Z","lastTransitionTime":"2025-12-06T13:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.473574 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.473612 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.473623 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.473639 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.473651 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:25Z","lastTransitionTime":"2025-12-06T13:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.511194 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.511230 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:25 crc kubenswrapper[4706]: E1206 13:56:25.511475 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:25 crc kubenswrapper[4706]: E1206 13:56:25.511771 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.578695 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.578749 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.578766 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.578793 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.578811 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:25Z","lastTransitionTime":"2025-12-06T13:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.682629 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.682701 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.682724 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.682747 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.682763 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:25Z","lastTransitionTime":"2025-12-06T13:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.783972 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerStarted","Data":"c4ca30f2d6f2ea5b50488e197b4134982bed7d766f750868209b401f9e03af3e"} Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.784075 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.784126 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.786617 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.786643 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.786653 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.786666 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.786677 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:25Z","lastTransitionTime":"2025-12-06T13:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.804624 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:25Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.822438 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.824491 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.838645 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ca30f2d6f2ea5b50488e197b4134982bed7d766f750868209b401f9e03af3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:25Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.858654 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:25Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.890051 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.890111 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.890128 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.890152 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.890171 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:25Z","lastTransitionTime":"2025-12-06T13:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.897963 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:25Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.915101 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:25Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.932578 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:25Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.948398 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:25Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.962027 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:25Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.974880 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:25Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.993283 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.993398 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.993423 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.993452 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.993476 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:25Z","lastTransitionTime":"2025-12-06T13:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:25 crc kubenswrapper[4706]: I1206 13:56:25.994656 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:25Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.013636 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.031745 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.052678 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.065958 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.084801 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.095748 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.095818 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.095848 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.095879 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.095901 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:26Z","lastTransitionTime":"2025-12-06T13:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.104079 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.125069 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.134001 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.134061 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.134079 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.134104 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.134124 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:26Z","lastTransitionTime":"2025-12-06T13:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.144262 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: E1206 13:56:26.154089 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.164977 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.164880 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.165844 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.165912 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.165998 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.166023 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:26Z","lastTransitionTime":"2025-12-06T13:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.182494 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: E1206 13:56:26.187268 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.193392 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.193448 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.193468 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.193493 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.193511 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:26Z","lastTransitionTime":"2025-12-06T13:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.212964 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ca30f2d6f2ea5b50488e197b4134982bed7d766f750868209b401f9e03af3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: E1206 13:56:26.217188 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.224248 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.224357 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.224378 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.224403 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.224421 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:26Z","lastTransitionTime":"2025-12-06T13:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.234566 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: E1206 13:56:26.245172 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.250143 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.250207 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.250225 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.250252 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.250272 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:26Z","lastTransitionTime":"2025-12-06T13:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:26 crc kubenswrapper[4706]: E1206 13:56:26.270533 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.270587 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: E1206 13:56:26.270755 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.273538 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.273594 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.273615 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.273640 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.273658 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:26Z","lastTransitionTime":"2025-12-06T13:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.289916 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.307194 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.328685 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.343875 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.365419 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.376520 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.376566 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.376583 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.376605 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.376622 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:26Z","lastTransitionTime":"2025-12-06T13:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.380257 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.402981 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:26Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.479768 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.479862 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.479881 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.479904 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.479921 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:26Z","lastTransitionTime":"2025-12-06T13:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.510449 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:26 crc kubenswrapper[4706]: E1206 13:56:26.510602 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.583401 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.583481 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.583504 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.583533 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.583554 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:26Z","lastTransitionTime":"2025-12-06T13:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.686013 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.686075 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.686091 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.686113 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.686134 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:26Z","lastTransitionTime":"2025-12-06T13:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.786373 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.788757 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.788806 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.788825 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.788848 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.788865 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:26Z","lastTransitionTime":"2025-12-06T13:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.892538 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.892606 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.892624 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.892648 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.892665 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:26Z","lastTransitionTime":"2025-12-06T13:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.996140 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.996186 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.996197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.996214 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:26 crc kubenswrapper[4706]: I1206 13:56:26.996229 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:26Z","lastTransitionTime":"2025-12-06T13:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.098370 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.098435 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.098454 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.098484 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.098503 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:27Z","lastTransitionTime":"2025-12-06T13:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.201129 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.201188 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.201205 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.201229 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.201247 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:27Z","lastTransitionTime":"2025-12-06T13:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.304554 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.304596 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.304606 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.304645 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.304657 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:27Z","lastTransitionTime":"2025-12-06T13:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.407977 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.408030 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.408046 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.408071 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.408089 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:27Z","lastTransitionTime":"2025-12-06T13:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.510721 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.514049 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:27 crc kubenswrapper[4706]: E1206 13:56:27.514238 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:27 crc kubenswrapper[4706]: E1206 13:56:27.514713 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.517927 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.517966 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.517985 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.518018 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.518995 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:27Z","lastTransitionTime":"2025-12-06T13:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.537885 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:27Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.553887 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:27Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.569009 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:27Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.598142 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ca30f2d6f2ea5b50488e197b4134982bed7d766f750868209b401f9e03af3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:27Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.624395 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.624452 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.624465 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.624481 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.624520 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:27Z","lastTransitionTime":"2025-12-06T13:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.626515 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:27Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.660008 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:27Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.679744 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:27Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.700474 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:27Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.722088 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:27Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.727455 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.727487 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.727496 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.727510 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.727519 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:27Z","lastTransitionTime":"2025-12-06T13:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.739860 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:27Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.761018 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:27Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.780235 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:27Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.790269 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.801450 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:27Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.828604 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:27Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.831179 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.831493 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.831635 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.831799 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.831996 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:27Z","lastTransitionTime":"2025-12-06T13:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.845809 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:27Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.934597 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.934670 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.934694 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.934723 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:27 crc kubenswrapper[4706]: I1206 13:56:27.934746 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:27Z","lastTransitionTime":"2025-12-06T13:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.036827 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.036885 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.036903 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.036926 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.036943 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:28Z","lastTransitionTime":"2025-12-06T13:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.139843 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.139891 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.139908 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.139933 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.139950 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:28Z","lastTransitionTime":"2025-12-06T13:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.243161 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.243216 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.243233 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.243257 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.243274 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:28Z","lastTransitionTime":"2025-12-06T13:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.346253 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.346314 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.346355 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.346378 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.346396 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:28Z","lastTransitionTime":"2025-12-06T13:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.449707 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.449750 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.449762 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.449779 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.449789 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:28Z","lastTransitionTime":"2025-12-06T13:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.510009 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:28 crc kubenswrapper[4706]: E1206 13:56:28.510181 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.550280 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.552177 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.552244 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.552272 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.552302 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.552377 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:28Z","lastTransitionTime":"2025-12-06T13:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.571604 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.589068 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.609093 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.626251 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.640647 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.654791 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.654843 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.654860 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.654884 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.654904 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:28Z","lastTransitionTime":"2025-12-06T13:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.665207 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ca30f2d6f2ea5b50488e197b4134982bed7d766f750868209b401f9e03af3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.683734 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.708812 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.721493 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.741216 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.758274 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.758394 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.758421 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.758452 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.758476 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:28Z","lastTransitionTime":"2025-12-06T13:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.760436 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.780137 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.796430 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovnkube-controller/0.log" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.800367 4706 generic.go:334] "Generic (PLEG): container finished" podID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerID="c4ca30f2d6f2ea5b50488e197b4134982bed7d766f750868209b401f9e03af3e" exitCode=1 Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.800422 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerDied","Data":"c4ca30f2d6f2ea5b50488e197b4134982bed7d766f750868209b401f9e03af3e"} Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.801612 4706 scope.go:117] "RemoveContainer" containerID="c4ca30f2d6f2ea5b50488e197b4134982bed7d766f750868209b401f9e03af3e" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.805260 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.822802 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.844598 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.861480 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.861539 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.861561 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.861589 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.861610 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:28Z","lastTransitionTime":"2025-12-06T13:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.864179 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.922876 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ca30f2d6f2ea5b50488e197b4134982bed7d766f750868209b401f9e03af3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4ca30f2d6f2ea5b50488e197b4134982bed7d766f750868209b401f9e03af3e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"message\\\":\\\" 6034 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:28.392053 6034 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392179 6034 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:28.392140 6034 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392252 6034 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392212 6034 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392292 6034 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 13:56:28.392370 6034 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.949951 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.964172 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.964206 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.964218 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.964233 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.964245 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:28Z","lastTransitionTime":"2025-12-06T13:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.972055 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:28 crc kubenswrapper[4706]: I1206 13:56:28.989158 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:28Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.003713 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.014893 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.024694 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.038508 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.055348 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.066490 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.066529 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.066543 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.066562 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.066577 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:29Z","lastTransitionTime":"2025-12-06T13:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.069523 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.083967 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.096790 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.114391 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.123135 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.169289 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.169346 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.169358 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.169375 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.169387 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:29Z","lastTransitionTime":"2025-12-06T13:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.271239 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.271280 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.271291 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.271307 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.271322 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:29Z","lastTransitionTime":"2025-12-06T13:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.374337 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.374388 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.374400 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.374416 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.374426 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:29Z","lastTransitionTime":"2025-12-06T13:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.476592 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.476637 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.476647 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.476664 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.476676 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:29Z","lastTransitionTime":"2025-12-06T13:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.510970 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.511000 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:29 crc kubenswrapper[4706]: E1206 13:56:29.511157 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:29 crc kubenswrapper[4706]: E1206 13:56:29.511308 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.579037 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.579069 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.579077 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.579090 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.579098 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:29Z","lastTransitionTime":"2025-12-06T13:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.681158 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.681191 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.681199 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.681213 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.681222 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:29Z","lastTransitionTime":"2025-12-06T13:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.784087 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.784173 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.784186 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.784212 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.784227 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:29Z","lastTransitionTime":"2025-12-06T13:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.806856 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovnkube-controller/0.log" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.810957 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerStarted","Data":"617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd"} Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.811156 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.824810 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.851920 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.869400 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.882458 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.886705 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.886770 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.886789 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.886815 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.886834 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:29Z","lastTransitionTime":"2025-12-06T13:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.896853 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.908538 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.932776 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4ca30f2d6f2ea5b50488e197b4134982bed7d766f750868209b401f9e03af3e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"message\\\":\\\" 6034 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:28.392053 6034 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392179 6034 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:28.392140 6034 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392252 6034 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392212 6034 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392292 6034 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 13:56:28.392370 6034 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.950529 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.967309 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.986699 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.988778 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.988898 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.989002 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.989065 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:29 crc kubenswrapper[4706]: I1206 13:56:29.989150 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:29Z","lastTransitionTime":"2025-12-06T13:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.001710 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:29Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.018795 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.042727 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.057857 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.079886 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.093209 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.093294 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.093312 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.093383 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.093405 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:30Z","lastTransitionTime":"2025-12-06T13:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.196930 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.197010 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.197032 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.197071 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.197097 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:30Z","lastTransitionTime":"2025-12-06T13:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.300258 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.300572 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.300682 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.300785 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.300866 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:30Z","lastTransitionTime":"2025-12-06T13:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.403873 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.403930 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.403946 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.403987 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.404004 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:30Z","lastTransitionTime":"2025-12-06T13:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.448512 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7"] Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.449448 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.452282 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.452496 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.468028 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.491356 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4ca30f2d6f2ea5b50488e197b4134982bed7d766f750868209b401f9e03af3e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"message\\\":\\\" 6034 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:28.392053 6034 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392179 6034 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:28.392140 6034 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392252 6034 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392212 6034 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392292 6034 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 13:56:28.392370 6034 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.506813 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.507007 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.507046 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.507063 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.507088 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.507105 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:30Z","lastTransitionTime":"2025-12-06T13:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.510590 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:30 crc kubenswrapper[4706]: E1206 13:56:30.510916 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.532164 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.533924 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a34577b8-1ea8-4af6-8585-0d9710dd30ee-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-dzzz7\" (UID: \"a34577b8-1ea8-4af6-8585-0d9710dd30ee\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.534063 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g54pg\" (UniqueName: \"kubernetes.io/projected/a34577b8-1ea8-4af6-8585-0d9710dd30ee-kube-api-access-g54pg\") pod \"ovnkube-control-plane-749d76644c-dzzz7\" (UID: \"a34577b8-1ea8-4af6-8585-0d9710dd30ee\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.534173 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a34577b8-1ea8-4af6-8585-0d9710dd30ee-env-overrides\") pod \"ovnkube-control-plane-749d76644c-dzzz7\" (UID: \"a34577b8-1ea8-4af6-8585-0d9710dd30ee\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.534262 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a34577b8-1ea8-4af6-8585-0d9710dd30ee-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-dzzz7\" (UID: \"a34577b8-1ea8-4af6-8585-0d9710dd30ee\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.547212 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.560084 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.575384 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.591741 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.609823 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.609883 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.609902 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.609933 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.609958 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:30Z","lastTransitionTime":"2025-12-06T13:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.611143 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.631860 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.635144 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a34577b8-1ea8-4af6-8585-0d9710dd30ee-env-overrides\") pod \"ovnkube-control-plane-749d76644c-dzzz7\" (UID: \"a34577b8-1ea8-4af6-8585-0d9710dd30ee\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.635233 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a34577b8-1ea8-4af6-8585-0d9710dd30ee-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-dzzz7\" (UID: \"a34577b8-1ea8-4af6-8585-0d9710dd30ee\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.635293 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g54pg\" (UniqueName: \"kubernetes.io/projected/a34577b8-1ea8-4af6-8585-0d9710dd30ee-kube-api-access-g54pg\") pod \"ovnkube-control-plane-749d76644c-dzzz7\" (UID: \"a34577b8-1ea8-4af6-8585-0d9710dd30ee\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.635362 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a34577b8-1ea8-4af6-8585-0d9710dd30ee-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-dzzz7\" (UID: \"a34577b8-1ea8-4af6-8585-0d9710dd30ee\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.636493 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a34577b8-1ea8-4af6-8585-0d9710dd30ee-env-overrides\") pod \"ovnkube-control-plane-749d76644c-dzzz7\" (UID: \"a34577b8-1ea8-4af6-8585-0d9710dd30ee\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.636673 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a34577b8-1ea8-4af6-8585-0d9710dd30ee-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-dzzz7\" (UID: \"a34577b8-1ea8-4af6-8585-0d9710dd30ee\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.643225 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a34577b8-1ea8-4af6-8585-0d9710dd30ee-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-dzzz7\" (UID: \"a34577b8-1ea8-4af6-8585-0d9710dd30ee\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.646671 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.655439 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g54pg\" (UniqueName: \"kubernetes.io/projected/a34577b8-1ea8-4af6-8585-0d9710dd30ee-kube-api-access-g54pg\") pod \"ovnkube-control-plane-749d76644c-dzzz7\" (UID: \"a34577b8-1ea8-4af6-8585-0d9710dd30ee\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.668596 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.686157 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.702516 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.712842 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.713022 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.713089 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.713165 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.713228 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:30Z","lastTransitionTime":"2025-12-06T13:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.715420 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.731564 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.768782 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" Dec 06 13:56:30 crc kubenswrapper[4706]: W1206 13:56:30.781235 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda34577b8_1ea8_4af6_8585_0d9710dd30ee.slice/crio-ee354b752e69bd9efd70403b6123d8812c7ab462c8b1e04684a3be6eaa402fec WatchSource:0}: Error finding container ee354b752e69bd9efd70403b6123d8812c7ab462c8b1e04684a3be6eaa402fec: Status 404 returned error can't find the container with id ee354b752e69bd9efd70403b6123d8812c7ab462c8b1e04684a3be6eaa402fec Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.822867 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.822912 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.822923 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.822941 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.822954 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:30Z","lastTransitionTime":"2025-12-06T13:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.824115 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovnkube-controller/1.log" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.824862 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovnkube-controller/0.log" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.830589 4706 generic.go:334] "Generic (PLEG): container finished" podID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerID="617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd" exitCode=1 Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.830677 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerDied","Data":"617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd"} Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.830742 4706 scope.go:117] "RemoveContainer" containerID="c4ca30f2d6f2ea5b50488e197b4134982bed7d766f750868209b401f9e03af3e" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.831850 4706 scope.go:117] "RemoveContainer" containerID="617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd" Dec 06 13:56:30 crc kubenswrapper[4706]: E1206 13:56:30.832120 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-q6bgc_openshift-ovn-kubernetes(c6db0d03-fbc0-4b3b-b89d-282767c274be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.833030 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" event={"ID":"a34577b8-1ea8-4af6-8585-0d9710dd30ee","Type":"ContainerStarted","Data":"ee354b752e69bd9efd70403b6123d8812c7ab462c8b1e04684a3be6eaa402fec"} Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.845752 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.871260 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.891184 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.908512 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.921548 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.925124 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.925164 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.925173 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.925194 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.925212 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:30Z","lastTransitionTime":"2025-12-06T13:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.933763 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.960539 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4ca30f2d6f2ea5b50488e197b4134982bed7d766f750868209b401f9e03af3e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"message\\\":\\\" 6034 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:28.392053 6034 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392179 6034 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:28.392140 6034 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392252 6034 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392212 6034 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392292 6034 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 13:56:28.392370 6034 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"message\\\":\\\"ory.go:160\\\\nI1206 13:56:29.689017 6170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:29.689042 6170 factory.go:656] Stopping watch factory\\\\nI1206 13:56:29.689047 6170 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 13:56:29.689065 6170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:29.689075 6170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 13:56:29.689189 6170 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689457 6170 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689561 6170 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689596 6170 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:29.689629 6170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689690 6170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.973669 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.986512 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:30 crc kubenswrapper[4706]: I1206 13:56:30.999918 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:30Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.013448 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.026810 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.029023 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.029177 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.029295 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.029424 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.029566 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:31Z","lastTransitionTime":"2025-12-06T13:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.039430 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.055160 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.068932 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.084140 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.132355 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.132701 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.132888 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.133145 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.133478 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:31Z","lastTransitionTime":"2025-12-06T13:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.184406 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-hhkhk"] Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.184764 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:31 crc kubenswrapper[4706]: E1206 13:56:31.184814 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.207970 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.227155 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.236838 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.236869 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.236880 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.236896 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.236931 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:31Z","lastTransitionTime":"2025-12-06T13:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.243770 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.259384 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.288531 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4ca30f2d6f2ea5b50488e197b4134982bed7d766f750868209b401f9e03af3e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"message\\\":\\\" 6034 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:28.392053 6034 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392179 6034 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:28.392140 6034 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392252 6034 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392212 6034 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392292 6034 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 13:56:28.392370 6034 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"message\\\":\\\"ory.go:160\\\\nI1206 13:56:29.689017 6170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:29.689042 6170 factory.go:656] Stopping watch factory\\\\nI1206 13:56:29.689047 6170 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 13:56:29.689065 6170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:29.689075 6170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 13:56:29.689189 6170 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689457 6170 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689561 6170 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689596 6170 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:29.689629 6170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689690 6170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.306915 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.337184 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.339840 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.339870 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.339879 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.339894 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.339908 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:31Z","lastTransitionTime":"2025-12-06T13:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.340770 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swfnh\" (UniqueName: \"kubernetes.io/projected/5443fc82-2d54-43df-88da-043c2eb77238-kube-api-access-swfnh\") pod \"network-metrics-daemon-hhkhk\" (UID: \"5443fc82-2d54-43df-88da-043c2eb77238\") " pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.341019 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs\") pod \"network-metrics-daemon-hhkhk\" (UID: \"5443fc82-2d54-43df-88da-043c2eb77238\") " pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.352785 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.371211 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.385753 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.398913 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5443fc82-2d54-43df-88da-043c2eb77238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhkhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.414294 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.435454 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.441665 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swfnh\" (UniqueName: \"kubernetes.io/projected/5443fc82-2d54-43df-88da-043c2eb77238-kube-api-access-swfnh\") pod \"network-metrics-daemon-hhkhk\" (UID: \"5443fc82-2d54-43df-88da-043c2eb77238\") " pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.441777 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.441973 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.441984 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.441949 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs\") pod \"network-metrics-daemon-hhkhk\" (UID: \"5443fc82-2d54-43df-88da-043c2eb77238\") " pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.442007 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.442123 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:31Z","lastTransitionTime":"2025-12-06T13:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:31 crc kubenswrapper[4706]: E1206 13:56:31.442337 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 13:56:31 crc kubenswrapper[4706]: E1206 13:56:31.442463 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs podName:5443fc82-2d54-43df-88da-043c2eb77238 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:31.942444898 +0000 UTC m=+34.778237775 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs") pod "network-metrics-daemon-hhkhk" (UID: "5443fc82-2d54-43df-88da-043c2eb77238") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.446965 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.456692 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swfnh\" (UniqueName: \"kubernetes.io/projected/5443fc82-2d54-43df-88da-043c2eb77238-kube-api-access-swfnh\") pod \"network-metrics-daemon-hhkhk\" (UID: \"5443fc82-2d54-43df-88da-043c2eb77238\") " pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.458110 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.470712 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.481023 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:31Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.510788 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:31 crc kubenswrapper[4706]: E1206 13:56:31.510991 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.511209 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:31 crc kubenswrapper[4706]: E1206 13:56:31.511453 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.544384 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.544439 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.544451 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.544467 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.544501 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:31Z","lastTransitionTime":"2025-12-06T13:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.646892 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.647058 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.647077 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.647107 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.647124 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:31Z","lastTransitionTime":"2025-12-06T13:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.749343 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.749627 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.749759 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.749911 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.750033 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:31Z","lastTransitionTime":"2025-12-06T13:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.853067 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.853124 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.853142 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.853167 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.853188 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:31Z","lastTransitionTime":"2025-12-06T13:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.947769 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs\") pod \"network-metrics-daemon-hhkhk\" (UID: \"5443fc82-2d54-43df-88da-043c2eb77238\") " pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:31 crc kubenswrapper[4706]: E1206 13:56:31.947939 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 13:56:31 crc kubenswrapper[4706]: E1206 13:56:31.947991 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs podName:5443fc82-2d54-43df-88da-043c2eb77238 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:32.947974453 +0000 UTC m=+35.783767320 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs") pod "network-metrics-daemon-hhkhk" (UID: "5443fc82-2d54-43df-88da-043c2eb77238") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.955125 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.955180 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.955197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.955222 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:31 crc kubenswrapper[4706]: I1206 13:56:31.955238 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:31Z","lastTransitionTime":"2025-12-06T13:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.058978 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.059289 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.059297 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.059324 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.059332 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:32Z","lastTransitionTime":"2025-12-06T13:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.161968 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.162012 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.162024 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.162043 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.162055 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:32Z","lastTransitionTime":"2025-12-06T13:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.252272 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.252462 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:32 crc kubenswrapper[4706]: E1206 13:56:32.252522 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:56:48.252484557 +0000 UTC m=+51.088277444 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:56:32 crc kubenswrapper[4706]: E1206 13:56:32.252689 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 13:56:32 crc kubenswrapper[4706]: E1206 13:56:32.252813 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:48.252786855 +0000 UTC m=+51.088579752 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.253084 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.253183 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.253219 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:32 crc kubenswrapper[4706]: E1206 13:56:32.253274 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 13:56:32 crc kubenswrapper[4706]: E1206 13:56:32.253301 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 13:56:32 crc kubenswrapper[4706]: E1206 13:56:32.253315 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 13:56:32 crc kubenswrapper[4706]: E1206 13:56:32.253345 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:32 crc kubenswrapper[4706]: E1206 13:56:32.253405 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:48.253388601 +0000 UTC m=+51.089181488 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 13:56:32 crc kubenswrapper[4706]: E1206 13:56:32.253429 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:48.253418512 +0000 UTC m=+51.089211409 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:32 crc kubenswrapper[4706]: E1206 13:56:32.253445 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 13:56:32 crc kubenswrapper[4706]: E1206 13:56:32.253474 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 13:56:32 crc kubenswrapper[4706]: E1206 13:56:32.253496 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:32 crc kubenswrapper[4706]: E1206 13:56:32.253560 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:48.253544315 +0000 UTC m=+51.089337212 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.266601 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.266672 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.266695 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.266725 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.266748 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:32Z","lastTransitionTime":"2025-12-06T13:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.369146 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.369193 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.369210 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.369233 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.369251 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:32Z","lastTransitionTime":"2025-12-06T13:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.471782 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.471825 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.471836 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.471852 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.471863 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:32Z","lastTransitionTime":"2025-12-06T13:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.510546 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:32 crc kubenswrapper[4706]: E1206 13:56:32.510724 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.573751 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.573798 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.573814 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.573836 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.573853 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:32Z","lastTransitionTime":"2025-12-06T13:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.677160 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.677233 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.677255 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.677284 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.677337 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:32Z","lastTransitionTime":"2025-12-06T13:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.780452 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.780519 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.780540 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.780565 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.780588 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:32Z","lastTransitionTime":"2025-12-06T13:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.840751 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovnkube-controller/1.log" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.846041 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" event={"ID":"a34577b8-1ea8-4af6-8585-0d9710dd30ee","Type":"ContainerStarted","Data":"272817fdf8aaa69e4b8cfa577e833c41292c789adbfaead6c4eea56351619dec"} Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.846100 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" event={"ID":"a34577b8-1ea8-4af6-8585-0d9710dd30ee","Type":"ContainerStarted","Data":"b18177422ea481c7f4a00b417f08daac333e453eafde2445d257280559d50ae3"} Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.865761 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:32Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.882254 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:32Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.882617 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.882675 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.882687 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.882702 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.882714 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:32Z","lastTransitionTime":"2025-12-06T13:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.900348 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:32Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.915101 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:32Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.926864 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:32Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.955152 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4ca30f2d6f2ea5b50488e197b4134982bed7d766f750868209b401f9e03af3e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"message\\\":\\\" 6034 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:28.392053 6034 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392179 6034 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:28.392140 6034 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392252 6034 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392212 6034 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:28.392292 6034 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 13:56:28.392370 6034 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"message\\\":\\\"ory.go:160\\\\nI1206 13:56:29.689017 6170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:29.689042 6170 factory.go:656] Stopping watch factory\\\\nI1206 13:56:29.689047 6170 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 13:56:29.689065 6170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:29.689075 6170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 13:56:29.689189 6170 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689457 6170 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689561 6170 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689596 6170 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:29.689629 6170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689690 6170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:32Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.961445 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs\") pod \"network-metrics-daemon-hhkhk\" (UID: \"5443fc82-2d54-43df-88da-043c2eb77238\") " pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:32 crc kubenswrapper[4706]: E1206 13:56:32.961742 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 13:56:32 crc kubenswrapper[4706]: E1206 13:56:32.961861 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs podName:5443fc82-2d54-43df-88da-043c2eb77238 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:34.961830877 +0000 UTC m=+37.797623764 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs") pod "network-metrics-daemon-hhkhk" (UID: "5443fc82-2d54-43df-88da-043c2eb77238") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.976507 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:32Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.985831 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.985878 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.985898 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.985920 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.985940 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:32Z","lastTransitionTime":"2025-12-06T13:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:32 crc kubenswrapper[4706]: I1206 13:56:32.999480 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:32Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.018914 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:33Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.038458 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:33Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.058205 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:33Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.073759 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5443fc82-2d54-43df-88da-043c2eb77238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhkhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:33Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.087771 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.087808 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.087824 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.087847 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.087863 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:33Z","lastTransitionTime":"2025-12-06T13:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.091624 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:33Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.116947 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:33Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.130928 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:33Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.145840 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b18177422ea481c7f4a00b417f08daac333e453eafde2445d257280559d50ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://272817fdf8aaa69e4b8cfa577e833c41292c789adbfaead6c4eea56351619dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:33Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.164491 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:33Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.189700 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.189768 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.189792 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.189817 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.189835 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:33Z","lastTransitionTime":"2025-12-06T13:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.292578 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.292636 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.292656 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.292678 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.292694 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:33Z","lastTransitionTime":"2025-12-06T13:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.395096 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.395152 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.395168 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.395193 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.395212 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:33Z","lastTransitionTime":"2025-12-06T13:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.498509 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.498575 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.498593 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.498617 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.498636 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:33Z","lastTransitionTime":"2025-12-06T13:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.510565 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.510634 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.510575 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:33 crc kubenswrapper[4706]: E1206 13:56:33.510796 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:33 crc kubenswrapper[4706]: E1206 13:56:33.510967 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:56:33 crc kubenswrapper[4706]: E1206 13:56:33.511142 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.601645 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.601706 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.601725 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.601771 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.601789 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:33Z","lastTransitionTime":"2025-12-06T13:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.705163 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.705222 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.705242 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.705267 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.705291 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:33Z","lastTransitionTime":"2025-12-06T13:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.809425 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.809483 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.809497 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.809556 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.809575 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:33Z","lastTransitionTime":"2025-12-06T13:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.913026 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.913090 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.913107 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.913135 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:33 crc kubenswrapper[4706]: I1206 13:56:33.913152 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:33Z","lastTransitionTime":"2025-12-06T13:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.016356 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.016420 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.016438 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.016463 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.016483 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:34Z","lastTransitionTime":"2025-12-06T13:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.119033 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.119082 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.119093 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.119113 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.119126 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:34Z","lastTransitionTime":"2025-12-06T13:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.222046 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.222095 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.222105 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.222119 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.222128 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:34Z","lastTransitionTime":"2025-12-06T13:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.325293 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.325391 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.325410 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.325432 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.325449 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:34Z","lastTransitionTime":"2025-12-06T13:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.428226 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.428272 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.428285 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.428307 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.428341 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:34Z","lastTransitionTime":"2025-12-06T13:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.510076 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:34 crc kubenswrapper[4706]: E1206 13:56:34.510248 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.531123 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.531181 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.531202 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.531232 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.531254 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:34Z","lastTransitionTime":"2025-12-06T13:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.634574 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.634634 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.634652 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.634674 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.634691 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:34Z","lastTransitionTime":"2025-12-06T13:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.737918 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.737989 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.738012 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.738043 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.738068 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:34Z","lastTransitionTime":"2025-12-06T13:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.841494 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.841575 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.841593 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.841619 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.841638 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:34Z","lastTransitionTime":"2025-12-06T13:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.944914 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.944978 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.944994 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.945018 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.945034 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:34Z","lastTransitionTime":"2025-12-06T13:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:34 crc kubenswrapper[4706]: I1206 13:56:34.983429 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs\") pod \"network-metrics-daemon-hhkhk\" (UID: \"5443fc82-2d54-43df-88da-043c2eb77238\") " pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:34 crc kubenswrapper[4706]: E1206 13:56:34.983616 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 13:56:34 crc kubenswrapper[4706]: E1206 13:56:34.983736 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs podName:5443fc82-2d54-43df-88da-043c2eb77238 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:38.983703533 +0000 UTC m=+41.819496430 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs") pod "network-metrics-daemon-hhkhk" (UID: "5443fc82-2d54-43df-88da-043c2eb77238") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.048237 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.048359 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.048380 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.048406 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.048426 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:35Z","lastTransitionTime":"2025-12-06T13:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.151165 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.151214 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.151232 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.151251 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.151263 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:35Z","lastTransitionTime":"2025-12-06T13:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.254161 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.254229 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.254249 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.254269 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.254285 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:35Z","lastTransitionTime":"2025-12-06T13:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.357226 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.357288 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.357304 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.357368 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.357391 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:35Z","lastTransitionTime":"2025-12-06T13:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.405348 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.406540 4706 scope.go:117] "RemoveContainer" containerID="617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd" Dec 06 13:56:35 crc kubenswrapper[4706]: E1206 13:56:35.406850 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-q6bgc_openshift-ovn-kubernetes(c6db0d03-fbc0-4b3b-b89d-282767c274be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.423482 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:35Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.454034 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"message\\\":\\\"ory.go:160\\\\nI1206 13:56:29.689017 6170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:29.689042 6170 factory.go:656] Stopping watch factory\\\\nI1206 13:56:29.689047 6170 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 13:56:29.689065 6170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:29.689075 6170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 13:56:29.689189 6170 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689457 6170 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689561 6170 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689596 6170 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:29.689629 6170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689690 6170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-q6bgc_openshift-ovn-kubernetes(c6db0d03-fbc0-4b3b-b89d-282767c274be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:35Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.459806 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.459871 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.459889 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.459914 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.459932 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:35Z","lastTransitionTime":"2025-12-06T13:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.475641 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:35Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.509137 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:35Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.510129 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.510153 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.510183 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:35 crc kubenswrapper[4706]: E1206 13:56:35.510388 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:56:35 crc kubenswrapper[4706]: E1206 13:56:35.510553 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:35 crc kubenswrapper[4706]: E1206 13:56:35.510688 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.531182 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:35Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.550935 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:35Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.563407 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.563462 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.563479 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.563502 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.563521 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:35Z","lastTransitionTime":"2025-12-06T13:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.571395 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:35Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.589606 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5443fc82-2d54-43df-88da-043c2eb77238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhkhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:35Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.609737 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:35Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.629579 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:35Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.650096 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:35Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.666842 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.666930 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.666950 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.666974 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.667031 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:35Z","lastTransitionTime":"2025-12-06T13:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.670599 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:35Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.694663 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:35Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.720467 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:35Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.735504 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:35Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.752015 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b18177422ea481c7f4a00b417f08daac333e453eafde2445d257280559d50ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://272817fdf8aaa69e4b8cfa577e833c41292c789adbfaead6c4eea56351619dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:35Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.771181 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:35Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.772575 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.772632 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.772650 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.772678 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.772696 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:35Z","lastTransitionTime":"2025-12-06T13:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.875514 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.875572 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.875591 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.875613 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.875632 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:35Z","lastTransitionTime":"2025-12-06T13:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.978692 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.978751 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.978767 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.978790 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:35 crc kubenswrapper[4706]: I1206 13:56:35.978807 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:35Z","lastTransitionTime":"2025-12-06T13:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.081929 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.082003 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.082021 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.082047 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.082066 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:36Z","lastTransitionTime":"2025-12-06T13:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.184793 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.184858 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.184874 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.184899 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.184917 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:36Z","lastTransitionTime":"2025-12-06T13:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.288141 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.288202 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.288221 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.288243 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.288261 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:36Z","lastTransitionTime":"2025-12-06T13:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.392083 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.392151 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.392171 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.392199 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.392235 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:36Z","lastTransitionTime":"2025-12-06T13:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.495125 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.495187 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.495205 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.495227 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.495245 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:36Z","lastTransitionTime":"2025-12-06T13:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.510917 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:36 crc kubenswrapper[4706]: E1206 13:56:36.511108 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.597888 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.597950 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.597966 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.597991 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.598008 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:36Z","lastTransitionTime":"2025-12-06T13:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.606110 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.606177 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.606193 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.606218 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.606237 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:36Z","lastTransitionTime":"2025-12-06T13:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:36 crc kubenswrapper[4706]: E1206 13:56:36.633174 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:36Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.638660 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.638722 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.638745 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.638773 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.638790 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:36Z","lastTransitionTime":"2025-12-06T13:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:36 crc kubenswrapper[4706]: E1206 13:56:36.659240 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:36Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.664436 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.664535 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.664555 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.664577 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.664594 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:36Z","lastTransitionTime":"2025-12-06T13:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:36 crc kubenswrapper[4706]: E1206 13:56:36.684240 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:36Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.689534 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.689593 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.689609 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.689634 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.689651 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:36Z","lastTransitionTime":"2025-12-06T13:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:36 crc kubenswrapper[4706]: E1206 13:56:36.715847 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:36Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.721760 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.721844 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.721869 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.721899 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.721921 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:36Z","lastTransitionTime":"2025-12-06T13:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:36 crc kubenswrapper[4706]: E1206 13:56:36.746480 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:36Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:36 crc kubenswrapper[4706]: E1206 13:56:36.746725 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.748846 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.748894 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.748910 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.748931 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.748948 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:36Z","lastTransitionTime":"2025-12-06T13:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.852559 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.852615 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.852632 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.852656 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.852673 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:36Z","lastTransitionTime":"2025-12-06T13:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.955551 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.955628 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.955713 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.955749 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:36 crc kubenswrapper[4706]: I1206 13:56:36.955772 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:36Z","lastTransitionTime":"2025-12-06T13:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.058638 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.058687 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.058703 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.058726 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.058745 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:37Z","lastTransitionTime":"2025-12-06T13:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.161939 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.162019 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.162045 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.162077 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.162098 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:37Z","lastTransitionTime":"2025-12-06T13:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.264608 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.264675 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.264708 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.264737 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.264756 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:37Z","lastTransitionTime":"2025-12-06T13:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.368124 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.368178 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.368194 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.368217 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.368233 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:37Z","lastTransitionTime":"2025-12-06T13:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.471620 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.471700 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.471720 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.471744 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.471763 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:37Z","lastTransitionTime":"2025-12-06T13:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.510734 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.510890 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:37 crc kubenswrapper[4706]: E1206 13:56:37.511030 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.511082 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:37 crc kubenswrapper[4706]: E1206 13:56:37.511203 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:56:37 crc kubenswrapper[4706]: E1206 13:56:37.511460 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.532997 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:37Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.549039 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5443fc82-2d54-43df-88da-043c2eb77238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhkhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:37Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.567802 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:37Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.574357 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.574399 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.574416 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.574441 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.574460 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:37Z","lastTransitionTime":"2025-12-06T13:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.585344 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:37Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.607987 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:37Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.628436 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b18177422ea481c7f4a00b417f08daac333e453eafde2445d257280559d50ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://272817fdf8aaa69e4b8cfa577e833c41292c789adbfaead6c4eea56351619dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:37Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.649981 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:37Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.673699 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:37Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.678049 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.678111 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.678133 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.678184 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.678209 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:37Z","lastTransitionTime":"2025-12-06T13:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.690894 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:37Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.710669 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:37Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.730065 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:37Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.747649 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:37Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.780996 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.781053 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.781071 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.781097 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.781115 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:37Z","lastTransitionTime":"2025-12-06T13:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.782438 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"message\\\":\\\"ory.go:160\\\\nI1206 13:56:29.689017 6170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:29.689042 6170 factory.go:656] Stopping watch factory\\\\nI1206 13:56:29.689047 6170 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 13:56:29.689065 6170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:29.689075 6170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 13:56:29.689189 6170 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689457 6170 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689561 6170 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689596 6170 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:29.689629 6170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689690 6170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-q6bgc_openshift-ovn-kubernetes(c6db0d03-fbc0-4b3b-b89d-282767c274be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:37Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.804645 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:37Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.873847 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:37Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.883711 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.883758 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.883775 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.883797 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.883813 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:37Z","lastTransitionTime":"2025-12-06T13:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.889411 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:37Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.905043 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:37Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.987199 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.987330 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.987350 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.987373 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:37 crc kubenswrapper[4706]: I1206 13:56:37.987393 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:37Z","lastTransitionTime":"2025-12-06T13:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.090261 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.090347 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.090366 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.090389 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.090407 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:38Z","lastTransitionTime":"2025-12-06T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.193152 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.193208 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.193224 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.193247 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.193265 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:38Z","lastTransitionTime":"2025-12-06T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.296397 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.296456 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.296474 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.296527 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.296544 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:38Z","lastTransitionTime":"2025-12-06T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.399834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.399898 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.399917 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.399941 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.399959 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:38Z","lastTransitionTime":"2025-12-06T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.503582 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.503658 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.503682 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.503712 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.503739 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:38Z","lastTransitionTime":"2025-12-06T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.510071 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:38 crc kubenswrapper[4706]: E1206 13:56:38.510242 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.606041 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.606102 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.606118 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.606148 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.606166 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:38Z","lastTransitionTime":"2025-12-06T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.709158 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.709220 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.709232 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.709249 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.709266 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:38Z","lastTransitionTime":"2025-12-06T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.813557 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.813622 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.813640 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.813666 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.813684 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:38Z","lastTransitionTime":"2025-12-06T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.917417 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.917490 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.917508 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.917534 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:38 crc kubenswrapper[4706]: I1206 13:56:38.917552 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:38Z","lastTransitionTime":"2025-12-06T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.020760 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.020831 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.020848 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.020870 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.020891 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:39Z","lastTransitionTime":"2025-12-06T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.031714 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs\") pod \"network-metrics-daemon-hhkhk\" (UID: \"5443fc82-2d54-43df-88da-043c2eb77238\") " pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:39 crc kubenswrapper[4706]: E1206 13:56:39.031896 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 13:56:39 crc kubenswrapper[4706]: E1206 13:56:39.032007 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs podName:5443fc82-2d54-43df-88da-043c2eb77238 nodeName:}" failed. No retries permitted until 2025-12-06 13:56:47.03197722 +0000 UTC m=+49.867770107 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs") pod "network-metrics-daemon-hhkhk" (UID: "5443fc82-2d54-43df-88da-043c2eb77238") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.124271 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.124394 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.124414 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.124472 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.124493 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:39Z","lastTransitionTime":"2025-12-06T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.229035 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.229109 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.229135 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.229169 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.229192 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:39Z","lastTransitionTime":"2025-12-06T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.332714 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.332822 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.332841 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.332898 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.332919 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:39Z","lastTransitionTime":"2025-12-06T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.436379 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.436438 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.436455 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.436479 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.436496 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:39Z","lastTransitionTime":"2025-12-06T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.510474 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.510500 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:39 crc kubenswrapper[4706]: E1206 13:56:39.510683 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.510722 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:39 crc kubenswrapper[4706]: E1206 13:56:39.510900 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:39 crc kubenswrapper[4706]: E1206 13:56:39.511067 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.539658 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.539707 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.539724 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.539749 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.539767 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:39Z","lastTransitionTime":"2025-12-06T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.642895 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.642963 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.642982 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.643009 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.643029 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:39Z","lastTransitionTime":"2025-12-06T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.746304 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.746393 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.746408 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.746430 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.746448 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:39Z","lastTransitionTime":"2025-12-06T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.849687 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.849739 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.849755 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.849779 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.849796 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:39Z","lastTransitionTime":"2025-12-06T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.953654 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.953719 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.953736 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.953762 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:39 crc kubenswrapper[4706]: I1206 13:56:39.953780 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:39Z","lastTransitionTime":"2025-12-06T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.057676 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.057739 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.057756 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.057779 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.057798 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:40Z","lastTransitionTime":"2025-12-06T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.161016 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.161079 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.161098 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.161124 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.161141 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:40Z","lastTransitionTime":"2025-12-06T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.264858 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.264925 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.264944 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.264969 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.264990 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:40Z","lastTransitionTime":"2025-12-06T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.368028 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.368184 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.368217 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.368249 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.368272 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:40Z","lastTransitionTime":"2025-12-06T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.471537 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.471598 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.471616 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.471643 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.471666 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:40Z","lastTransitionTime":"2025-12-06T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.510395 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:40 crc kubenswrapper[4706]: E1206 13:56:40.510625 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.574680 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.574749 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.574766 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.574792 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.574812 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:40Z","lastTransitionTime":"2025-12-06T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.677688 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.677776 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.677799 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.677830 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.677852 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:40Z","lastTransitionTime":"2025-12-06T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.780852 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.780940 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.780965 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.780998 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.781020 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:40Z","lastTransitionTime":"2025-12-06T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.883785 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.883864 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.883885 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.883913 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.883931 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:40Z","lastTransitionTime":"2025-12-06T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.987180 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.987257 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.987276 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.987341 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:40 crc kubenswrapper[4706]: I1206 13:56:40.987363 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:40Z","lastTransitionTime":"2025-12-06T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.090996 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.091062 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.091078 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.091103 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.091121 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:41Z","lastTransitionTime":"2025-12-06T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.194409 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.194474 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.194492 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.194521 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.194542 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:41Z","lastTransitionTime":"2025-12-06T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.297152 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.297213 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.297230 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.297254 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.297274 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:41Z","lastTransitionTime":"2025-12-06T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.399646 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.399741 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.399759 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.399784 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.399802 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:41Z","lastTransitionTime":"2025-12-06T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.503423 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.503484 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.503500 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.503522 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.503540 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:41Z","lastTransitionTime":"2025-12-06T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.510971 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.511089 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.510985 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:41 crc kubenswrapper[4706]: E1206 13:56:41.511276 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:56:41 crc kubenswrapper[4706]: E1206 13:56:41.511154 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:41 crc kubenswrapper[4706]: E1206 13:56:41.511665 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.607081 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.607138 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.607157 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.607182 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.607203 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:41Z","lastTransitionTime":"2025-12-06T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.710162 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.710235 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.710255 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.710284 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.710304 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:41Z","lastTransitionTime":"2025-12-06T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.814405 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.814469 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.814483 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.814512 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.814539 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:41Z","lastTransitionTime":"2025-12-06T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.918214 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.918295 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.918359 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.918394 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:41 crc kubenswrapper[4706]: I1206 13:56:41.918421 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:41Z","lastTransitionTime":"2025-12-06T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.022516 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.022564 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.022577 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.022597 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.022610 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:42Z","lastTransitionTime":"2025-12-06T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.126197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.126271 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.126291 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.126338 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.126357 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:42Z","lastTransitionTime":"2025-12-06T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.229895 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.229962 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.229979 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.230004 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.230022 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:42Z","lastTransitionTime":"2025-12-06T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.333736 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.333807 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.333824 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.333853 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.333873 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:42Z","lastTransitionTime":"2025-12-06T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.437493 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.437600 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.437620 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.437683 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.437701 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:42Z","lastTransitionTime":"2025-12-06T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.510597 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:42 crc kubenswrapper[4706]: E1206 13:56:42.511062 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.540761 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.541032 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.541229 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.541509 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.541677 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:42Z","lastTransitionTime":"2025-12-06T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.645412 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.645746 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.645896 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.646094 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.646381 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:42Z","lastTransitionTime":"2025-12-06T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.750120 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.750403 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.750569 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.750718 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.750863 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:42Z","lastTransitionTime":"2025-12-06T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.854861 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.855280 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.855804 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.856165 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.856616 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:42Z","lastTransitionTime":"2025-12-06T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.960806 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.960908 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.960927 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.960961 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:42 crc kubenswrapper[4706]: I1206 13:56:42.960980 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:42Z","lastTransitionTime":"2025-12-06T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.064728 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.064779 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.064795 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.064821 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.064841 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:43Z","lastTransitionTime":"2025-12-06T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.167444 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.167507 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.167523 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.167548 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.167566 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:43Z","lastTransitionTime":"2025-12-06T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.271033 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.271990 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.272123 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.272256 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.272452 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:43Z","lastTransitionTime":"2025-12-06T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.375359 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.375723 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.375869 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.376009 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.376127 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:43Z","lastTransitionTime":"2025-12-06T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.479471 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.479538 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.479556 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.479584 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.479602 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:43Z","lastTransitionTime":"2025-12-06T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.510155 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.510205 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.510193 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:43 crc kubenswrapper[4706]: E1206 13:56:43.510605 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:43 crc kubenswrapper[4706]: E1206 13:56:43.510519 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:43 crc kubenswrapper[4706]: E1206 13:56:43.510970 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.584011 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.584075 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.584098 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.584128 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.584149 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:43Z","lastTransitionTime":"2025-12-06T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.687955 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.688031 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.688058 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.688088 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.688106 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:43Z","lastTransitionTime":"2025-12-06T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.790838 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.790889 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.790909 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.790940 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.790962 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:43Z","lastTransitionTime":"2025-12-06T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.894014 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.894081 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.894103 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.894135 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.894158 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:43Z","lastTransitionTime":"2025-12-06T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.996684 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.996867 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.996895 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.996924 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:43 crc kubenswrapper[4706]: I1206 13:56:43.996950 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:43Z","lastTransitionTime":"2025-12-06T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.100179 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.100263 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.100420 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.100512 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.100540 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:44Z","lastTransitionTime":"2025-12-06T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.207764 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.207820 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.207851 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.207877 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.207897 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:44Z","lastTransitionTime":"2025-12-06T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.312015 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.313501 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.313810 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.314019 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.314404 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:44Z","lastTransitionTime":"2025-12-06T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.417483 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.417832 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.417989 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.418293 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.418490 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:44Z","lastTransitionTime":"2025-12-06T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.510225 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:44 crc kubenswrapper[4706]: E1206 13:56:44.510489 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.521488 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.521696 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.521849 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.521983 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.522131 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:44Z","lastTransitionTime":"2025-12-06T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.625500 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.625679 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.625711 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.625740 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.625764 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:44Z","lastTransitionTime":"2025-12-06T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.728766 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.729120 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.729256 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.729470 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.729625 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:44Z","lastTransitionTime":"2025-12-06T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.832184 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.832240 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.832257 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.832281 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.832297 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:44Z","lastTransitionTime":"2025-12-06T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.934601 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.934657 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.934675 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.934699 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:44 crc kubenswrapper[4706]: I1206 13:56:44.934721 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:44Z","lastTransitionTime":"2025-12-06T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.037721 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.037790 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.037807 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.037832 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.037852 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:45Z","lastTransitionTime":"2025-12-06T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.141092 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.141562 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.141730 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.141885 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.142041 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:45Z","lastTransitionTime":"2025-12-06T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.245056 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.245115 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.245132 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.245170 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.245188 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:45Z","lastTransitionTime":"2025-12-06T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.348476 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.348538 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.348558 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.348585 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.348629 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:45Z","lastTransitionTime":"2025-12-06T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.451359 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.451417 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.451435 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.451458 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.451477 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:45Z","lastTransitionTime":"2025-12-06T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.510673 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.510727 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.510841 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:45 crc kubenswrapper[4706]: E1206 13:56:45.510832 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:45 crc kubenswrapper[4706]: E1206 13:56:45.510989 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:56:45 crc kubenswrapper[4706]: E1206 13:56:45.511141 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.553961 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.554025 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.554050 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.554078 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.554100 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:45Z","lastTransitionTime":"2025-12-06T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.657012 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.657089 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.657107 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.657131 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.657149 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:45Z","lastTransitionTime":"2025-12-06T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.760378 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.760448 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.760467 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.760491 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.760509 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:45Z","lastTransitionTime":"2025-12-06T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.863424 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.863487 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.863503 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.863527 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.863544 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:45Z","lastTransitionTime":"2025-12-06T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.966059 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.966132 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.966157 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.966279 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:45 crc kubenswrapper[4706]: I1206 13:56:45.966339 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:45Z","lastTransitionTime":"2025-12-06T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.069802 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.069878 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.069901 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.069927 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.069944 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:46Z","lastTransitionTime":"2025-12-06T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.173212 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.173290 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.173349 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.173382 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.173406 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:46Z","lastTransitionTime":"2025-12-06T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.277125 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.277224 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.277246 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.277271 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.277289 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:46Z","lastTransitionTime":"2025-12-06T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.381542 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.381623 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.381669 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.381705 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.381728 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:46Z","lastTransitionTime":"2025-12-06T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.485039 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.485116 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.485136 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.485162 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.485181 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:46Z","lastTransitionTime":"2025-12-06T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.510692 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:46 crc kubenswrapper[4706]: E1206 13:56:46.510888 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.512114 4706 scope.go:117] "RemoveContainer" containerID="617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.588337 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.588643 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.588663 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.588688 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.588706 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:46Z","lastTransitionTime":"2025-12-06T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.692146 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.692211 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.692229 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.692253 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.692270 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:46Z","lastTransitionTime":"2025-12-06T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.795171 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.795220 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.795243 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.795274 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.795295 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:46Z","lastTransitionTime":"2025-12-06T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.899540 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.899570 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.899578 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.899591 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.899600 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:46Z","lastTransitionTime":"2025-12-06T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.903578 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovnkube-controller/1.log" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.906450 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerStarted","Data":"dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1"} Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.906732 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.925288 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:46Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.941055 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:46Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.967746 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"message\\\":\\\"ory.go:160\\\\nI1206 13:56:29.689017 6170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:29.689042 6170 factory.go:656] Stopping watch factory\\\\nI1206 13:56:29.689047 6170 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 13:56:29.689065 6170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:29.689075 6170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 13:56:29.689189 6170 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689457 6170 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689561 6170 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689596 6170 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:29.689629 6170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689690 6170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:46Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:46 crc kubenswrapper[4706]: I1206 13:56:46.981600 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:46Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.001604 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.001670 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.001691 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.001717 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.001739 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:47Z","lastTransitionTime":"2025-12-06T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.014292 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.033178 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: E1206 13:56:47.033589 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.033441 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs\") pod \"network-metrics-daemon-hhkhk\" (UID: \"5443fc82-2d54-43df-88da-043c2eb77238\") " pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:47 crc kubenswrapper[4706]: E1206 13:56:47.033681 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs podName:5443fc82-2d54-43df-88da-043c2eb77238 nodeName:}" failed. No retries permitted until 2025-12-06 13:57:03.033652261 +0000 UTC m=+65.869445158 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs") pod "network-metrics-daemon-hhkhk" (UID: "5443fc82-2d54-43df-88da-043c2eb77238") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.054152 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.054207 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.054229 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.054259 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.054283 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:47Z","lastTransitionTime":"2025-12-06T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.058669 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.075967 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: E1206 13:56:47.077810 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.085252 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.085301 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.085358 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.085392 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.085412 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:47Z","lastTransitionTime":"2025-12-06T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.101264 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5443fc82-2d54-43df-88da-043c2eb77238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhkhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: E1206 13:56:47.106231 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.110716 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.110797 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.110821 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.110853 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.110876 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:47Z","lastTransitionTime":"2025-12-06T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:47 crc kubenswrapper[4706]: E1206 13:56:47.135649 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.146003 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.148179 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.148233 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.148250 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.148272 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.148290 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:47Z","lastTransitionTime":"2025-12-06T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.165100 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: E1206 13:56:47.166649 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.170136 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.170276 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.170417 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.170532 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.170624 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:47Z","lastTransitionTime":"2025-12-06T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.177433 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: E1206 13:56:47.184754 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: E1206 13:56:47.185132 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.186738 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.186872 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.186964 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.187066 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.187158 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:47Z","lastTransitionTime":"2025-12-06T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.191522 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.203813 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.217553 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.228549 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.239818 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b18177422ea481c7f4a00b417f08daac333e453eafde2445d257280559d50ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://272817fdf8aaa69e4b8cfa577e833c41292c789adbfaead6c4eea56351619dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.289612 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.289653 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.289662 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.289677 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.289687 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:47Z","lastTransitionTime":"2025-12-06T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.391643 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.391701 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.391717 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.391741 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.391757 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:47Z","lastTransitionTime":"2025-12-06T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.494609 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.494981 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.495110 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.495253 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.495447 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:47Z","lastTransitionTime":"2025-12-06T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.510174 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.510259 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:47 crc kubenswrapper[4706]: E1206 13:56:47.510437 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:47 crc kubenswrapper[4706]: E1206 13:56:47.510577 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.511288 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:47 crc kubenswrapper[4706]: E1206 13:56:47.511529 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.537436 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.570088 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.586640 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.598639 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.598702 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.598720 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.598748 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.598765 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:47Z","lastTransitionTime":"2025-12-06T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.604981 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.620142 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.635079 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.655682 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"message\\\":\\\"ory.go:160\\\\nI1206 13:56:29.689017 6170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:29.689042 6170 factory.go:656] Stopping watch factory\\\\nI1206 13:56:29.689047 6170 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 13:56:29.689065 6170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:29.689075 6170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 13:56:29.689189 6170 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689457 6170 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689561 6170 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689596 6170 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:29.689629 6170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689690 6170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.669591 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.684857 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.696571 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.701081 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.701117 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.701129 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.701150 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.701164 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:47Z","lastTransitionTime":"2025-12-06T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.710988 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.725277 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5443fc82-2d54-43df-88da-043c2eb77238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhkhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.745019 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.767377 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.782580 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.803840 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.803880 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.803892 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.803911 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.803925 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:47Z","lastTransitionTime":"2025-12-06T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.804841 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b18177422ea481c7f4a00b417f08daac333e453eafde2445d257280559d50ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://272817fdf8aaa69e4b8cfa577e833c41292c789adbfaead6c4eea56351619dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.832345 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:47Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.907899 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.907968 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.907987 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.908015 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:47 crc kubenswrapper[4706]: I1206 13:56:47.908036 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:47Z","lastTransitionTime":"2025-12-06T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.011493 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.011566 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.011584 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.011612 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.011632 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:48Z","lastTransitionTime":"2025-12-06T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.118445 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.118523 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.118550 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.118581 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.118606 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:48Z","lastTransitionTime":"2025-12-06T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.222416 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.222486 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.222504 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.222532 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.222552 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:48Z","lastTransitionTime":"2025-12-06T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.327051 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.327131 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.327149 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.327178 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.327197 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:48Z","lastTransitionTime":"2025-12-06T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.348521 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:56:48 crc kubenswrapper[4706]: E1206 13:56:48.348745 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:57:20.348699893 +0000 UTC m=+83.184492790 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.348831 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.348927 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.348993 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.349063 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:48 crc kubenswrapper[4706]: E1206 13:56:48.349118 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 13:56:48 crc kubenswrapper[4706]: E1206 13:56:48.349269 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 13:57:20.349229587 +0000 UTC m=+83.185022624 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 13:56:48 crc kubenswrapper[4706]: E1206 13:56:48.349290 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 13:56:48 crc kubenswrapper[4706]: E1206 13:56:48.349299 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 13:56:48 crc kubenswrapper[4706]: E1206 13:56:48.349408 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 13:56:48 crc kubenswrapper[4706]: E1206 13:56:48.349433 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:48 crc kubenswrapper[4706]: E1206 13:56:48.349510 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 13:56:48 crc kubenswrapper[4706]: E1206 13:56:48.349528 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 13:57:20.349494524 +0000 UTC m=+83.185287411 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:48 crc kubenswrapper[4706]: E1206 13:56:48.349361 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 13:56:48 crc kubenswrapper[4706]: E1206 13:56:48.349816 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:48 crc kubenswrapper[4706]: E1206 13:56:48.349846 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 13:57:20.34973516 +0000 UTC m=+83.185528057 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 13:56:48 crc kubenswrapper[4706]: E1206 13:56:48.349927 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 13:57:20.349888894 +0000 UTC m=+83.185681791 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.431254 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.431340 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.431358 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.431379 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.431395 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:48Z","lastTransitionTime":"2025-12-06T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.510823 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:48 crc kubenswrapper[4706]: E1206 13:56:48.511085 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.536129 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.536202 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.536220 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.536246 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.536268 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:48Z","lastTransitionTime":"2025-12-06T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.638802 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.638854 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.638865 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.638881 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.638892 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:48Z","lastTransitionTime":"2025-12-06T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.742290 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.742370 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.742386 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.742411 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.742428 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:48Z","lastTransitionTime":"2025-12-06T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.770261 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.784568 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.787878 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:48Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.804180 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:48Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.819786 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:48Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.832361 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b18177422ea481c7f4a00b417f08daac333e453eafde2445d257280559d50ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://272817fdf8aaa69e4b8cfa577e833c41292c789adbfaead6c4eea56351619dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:48Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.845653 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.845701 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.845714 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.845731 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.845742 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:48Z","lastTransitionTime":"2025-12-06T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.847072 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:48Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.861964 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:48Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.880776 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:48Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.896575 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:48Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.912676 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:48Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.919005 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovnkube-controller/2.log" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.919894 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovnkube-controller/1.log" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.923984 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:48Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.924898 4706 generic.go:334] "Generic (PLEG): container finished" podID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerID="dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1" exitCode=1 Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.924980 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerDied","Data":"dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1"} Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.925059 4706 scope.go:117] "RemoveContainer" containerID="617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.931027 4706 scope.go:117] "RemoveContainer" containerID="dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1" Dec 06 13:56:48 crc kubenswrapper[4706]: E1206 13:56:48.931183 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-q6bgc_openshift-ovn-kubernetes(c6db0d03-fbc0-4b3b-b89d-282767c274be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.942269 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:48Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.957094 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.957131 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.957147 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.957162 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.957174 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:48Z","lastTransitionTime":"2025-12-06T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.974790 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"message\\\":\\\"ory.go:160\\\\nI1206 13:56:29.689017 6170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:29.689042 6170 factory.go:656] Stopping watch factory\\\\nI1206 13:56:29.689047 6170 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 13:56:29.689065 6170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:29.689075 6170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 13:56:29.689189 6170 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689457 6170 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689561 6170 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689596 6170 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:29.689629 6170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689690 6170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:48Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:48 crc kubenswrapper[4706]: I1206 13:56:48.989279 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:48Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.006646 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.027536 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.038464 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.047260 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5443fc82-2d54-43df-88da-043c2eb77238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhkhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.059870 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.059921 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.059932 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.059947 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.059959 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:49Z","lastTransitionTime":"2025-12-06T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.065650 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.083183 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.095911 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.107821 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b18177422ea481c7f4a00b417f08daac333e453eafde2445d257280559d50ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://272817fdf8aaa69e4b8cfa577e833c41292c789adbfaead6c4eea56351619dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.119495 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.132545 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.160646 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"message\\\":\\\"ory.go:160\\\\nI1206 13:56:29.689017 6170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:29.689042 6170 factory.go:656] Stopping watch factory\\\\nI1206 13:56:29.689047 6170 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 13:56:29.689065 6170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:29.689075 6170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 13:56:29.689189 6170 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689457 6170 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689561 6170 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689596 6170 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:29.689629 6170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689690 6170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"\\\\nI1206 13:56:47.623888 6367 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:47.624014 6367 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:47.624046 6367 factory.go:656] Stopping watch factory\\\\nI1206 13:56:47.624072 6367 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:47.624143 6367 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:47.624422 6367 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624704 6367 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624870 6367 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624895 6367 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624934 6367 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.162582 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.162613 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.162626 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.162646 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.162658 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:49Z","lastTransitionTime":"2025-12-06T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.177352 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.198384 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.217986 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.234645 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.250251 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.264868 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.264941 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.264963 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.264992 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.265019 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:49Z","lastTransitionTime":"2025-12-06T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.271890 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5443fc82-2d54-43df-88da-043c2eb77238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhkhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.288509 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873f40ed-e71b-4cf8-9322-b5ddda4ec8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://631b8b4aeac64fbdd48083db4545f8df35fa2e62abb03fcfc0d78ee39216be3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0730aa973ef51ef8c4d240fde0ef6135c76ac40b576a1afa846e8b3772c16120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c44a305126e75b14cf63cc84ae01d0d04203db32abb5f4434f1339a8e773fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.306544 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.321626 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.336774 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.357449 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:49Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.367335 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.367394 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.367406 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.367424 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.367436 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:49Z","lastTransitionTime":"2025-12-06T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.470838 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.470913 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.470933 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.470963 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.470981 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:49Z","lastTransitionTime":"2025-12-06T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.510813 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.510895 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:49 crc kubenswrapper[4706]: E1206 13:56:49.511022 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.511050 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:49 crc kubenswrapper[4706]: E1206 13:56:49.511224 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:56:49 crc kubenswrapper[4706]: E1206 13:56:49.511381 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.573768 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.573841 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.573857 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.573883 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.573903 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:49Z","lastTransitionTime":"2025-12-06T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.677624 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.677685 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.677702 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.677729 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.677770 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:49Z","lastTransitionTime":"2025-12-06T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.781275 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.781580 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.781596 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.781621 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.781639 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:49Z","lastTransitionTime":"2025-12-06T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.884760 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.884834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.884851 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.884876 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.884899 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:49Z","lastTransitionTime":"2025-12-06T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.930686 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovnkube-controller/2.log" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.987137 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.987206 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.987223 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.987260 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:49 crc kubenswrapper[4706]: I1206 13:56:49.987281 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:49Z","lastTransitionTime":"2025-12-06T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.090560 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.090612 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.090633 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.090656 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.090673 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:50Z","lastTransitionTime":"2025-12-06T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.193262 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.193348 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.193365 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.193388 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.193409 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:50Z","lastTransitionTime":"2025-12-06T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.300857 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.300915 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.300931 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.300954 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.300970 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:50Z","lastTransitionTime":"2025-12-06T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.404370 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.404433 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.404452 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.404476 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.404494 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:50Z","lastTransitionTime":"2025-12-06T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.507969 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.508025 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.508042 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.508065 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.508083 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:50Z","lastTransitionTime":"2025-12-06T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.510583 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:50 crc kubenswrapper[4706]: E1206 13:56:50.510758 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.611022 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.611132 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.611171 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.611195 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.611212 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:50Z","lastTransitionTime":"2025-12-06T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.714491 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.714561 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.714585 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.714613 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.714632 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:50Z","lastTransitionTime":"2025-12-06T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.817075 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.817146 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.817170 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.817202 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.817223 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:50Z","lastTransitionTime":"2025-12-06T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.920625 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.920665 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.920679 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.920697 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:50 crc kubenswrapper[4706]: I1206 13:56:50.920709 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:50Z","lastTransitionTime":"2025-12-06T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.023489 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.023552 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.023569 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.023595 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.023615 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:51Z","lastTransitionTime":"2025-12-06T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.126791 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.126858 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.126875 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.126899 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.126917 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:51Z","lastTransitionTime":"2025-12-06T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.229218 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.229280 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.229298 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.229349 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.229367 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:51Z","lastTransitionTime":"2025-12-06T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.331782 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.331816 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.331827 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.331843 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.331854 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:51Z","lastTransitionTime":"2025-12-06T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.434067 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.434126 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.434144 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.434166 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.434184 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:51Z","lastTransitionTime":"2025-12-06T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.510821 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.510879 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:51 crc kubenswrapper[4706]: E1206 13:56:51.511036 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.511087 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:51 crc kubenswrapper[4706]: E1206 13:56:51.511171 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:51 crc kubenswrapper[4706]: E1206 13:56:51.511280 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.536577 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.536629 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.536647 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.536692 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.536713 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:51Z","lastTransitionTime":"2025-12-06T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.640060 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.640124 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.640141 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.640161 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.640177 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:51Z","lastTransitionTime":"2025-12-06T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.743254 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.743384 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.743411 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.743438 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.743456 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:51Z","lastTransitionTime":"2025-12-06T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.845741 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.845798 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.845814 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.845836 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.845851 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:51Z","lastTransitionTime":"2025-12-06T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.948413 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.948496 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.948521 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.948550 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:51 crc kubenswrapper[4706]: I1206 13:56:51.948576 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:51Z","lastTransitionTime":"2025-12-06T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.051495 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.051560 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.051583 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.051616 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.051640 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:52Z","lastTransitionTime":"2025-12-06T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.154525 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.154604 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.154622 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.154645 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.154667 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:52Z","lastTransitionTime":"2025-12-06T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.257649 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.257759 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.257784 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.257816 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.257839 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:52Z","lastTransitionTime":"2025-12-06T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.360930 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.361002 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.361020 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.361044 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.361062 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:52Z","lastTransitionTime":"2025-12-06T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.464463 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.464526 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.464545 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.464572 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.464596 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:52Z","lastTransitionTime":"2025-12-06T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.510806 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:52 crc kubenswrapper[4706]: E1206 13:56:52.510985 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.567914 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.567970 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.567991 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.568020 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.568042 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:52Z","lastTransitionTime":"2025-12-06T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.670559 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.670630 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.670653 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.670681 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.670697 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:52Z","lastTransitionTime":"2025-12-06T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.773813 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.773864 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.773880 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.773901 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.773917 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:52Z","lastTransitionTime":"2025-12-06T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.876980 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.877357 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.877565 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.877777 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.877960 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:52Z","lastTransitionTime":"2025-12-06T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.980892 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.981030 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.981054 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.981077 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:52 crc kubenswrapper[4706]: I1206 13:56:52.981095 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:52Z","lastTransitionTime":"2025-12-06T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.084383 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.084435 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.084452 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.084474 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.084493 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:53Z","lastTransitionTime":"2025-12-06T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.187931 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.187979 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.187997 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.188020 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.188038 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:53Z","lastTransitionTime":"2025-12-06T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.290488 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.290550 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.290570 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.290594 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.290612 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:53Z","lastTransitionTime":"2025-12-06T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.393473 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.393558 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.393579 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.393611 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.393632 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:53Z","lastTransitionTime":"2025-12-06T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.496209 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.496253 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.496266 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.496284 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.496296 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:53Z","lastTransitionTime":"2025-12-06T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.510752 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:53 crc kubenswrapper[4706]: E1206 13:56:53.510934 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.510777 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.510981 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:53 crc kubenswrapper[4706]: E1206 13:56:53.511054 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:53 crc kubenswrapper[4706]: E1206 13:56:53.511137 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.599291 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.599424 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.599450 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.599481 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.599503 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:53Z","lastTransitionTime":"2025-12-06T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.710097 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.710164 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.710176 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.710192 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.710204 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:53Z","lastTransitionTime":"2025-12-06T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.812543 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.812588 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.812604 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.812628 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.812646 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:53Z","lastTransitionTime":"2025-12-06T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.916523 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.916592 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.916610 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.916638 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:53 crc kubenswrapper[4706]: I1206 13:56:53.916655 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:53Z","lastTransitionTime":"2025-12-06T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.019205 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.019277 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.019304 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.019377 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.019399 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:54Z","lastTransitionTime":"2025-12-06T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.122699 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.122763 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.122780 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.122803 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.122823 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:54Z","lastTransitionTime":"2025-12-06T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.226803 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.226911 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.226936 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.226971 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.226994 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:54Z","lastTransitionTime":"2025-12-06T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.330144 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.330196 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.330205 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.330220 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.330234 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:54Z","lastTransitionTime":"2025-12-06T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.432746 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.432809 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.432821 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.432843 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.432858 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:54Z","lastTransitionTime":"2025-12-06T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.510120 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:54 crc kubenswrapper[4706]: E1206 13:56:54.510272 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.535399 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.536174 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.536369 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.536571 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.536715 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:54Z","lastTransitionTime":"2025-12-06T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.639674 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.639746 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.639764 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.639790 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.639809 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:54Z","lastTransitionTime":"2025-12-06T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.742489 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.742934 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.743048 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.743158 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.743244 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:54Z","lastTransitionTime":"2025-12-06T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.846563 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.846627 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.846641 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.846664 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.846682 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:54Z","lastTransitionTime":"2025-12-06T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.951010 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.951081 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.951098 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.951122 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:54 crc kubenswrapper[4706]: I1206 13:56:54.951139 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:54Z","lastTransitionTime":"2025-12-06T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.054261 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.054415 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.054435 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.054460 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.054477 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:55Z","lastTransitionTime":"2025-12-06T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.157545 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.157604 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.157622 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.157646 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.157663 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:55Z","lastTransitionTime":"2025-12-06T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.260473 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.260535 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.260552 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.260577 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.260598 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:55Z","lastTransitionTime":"2025-12-06T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.364026 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.364107 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.364140 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.364168 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.364193 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:55Z","lastTransitionTime":"2025-12-06T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.467460 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.467530 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.467549 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.467575 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.467593 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:55Z","lastTransitionTime":"2025-12-06T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.510770 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.510820 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:55 crc kubenswrapper[4706]: E1206 13:56:55.510962 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.511045 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:55 crc kubenswrapper[4706]: E1206 13:56:55.511232 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:55 crc kubenswrapper[4706]: E1206 13:56:55.511364 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.570788 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.570869 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.570892 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.570922 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.570943 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:55Z","lastTransitionTime":"2025-12-06T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.674298 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.674660 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.674812 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.674967 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.675119 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:55Z","lastTransitionTime":"2025-12-06T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.779224 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.779264 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.779274 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.779289 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.779300 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:55Z","lastTransitionTime":"2025-12-06T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.881944 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.882031 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.882050 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.882074 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.882092 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:55Z","lastTransitionTime":"2025-12-06T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.984714 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.984821 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.984839 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.984862 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:55 crc kubenswrapper[4706]: I1206 13:56:55.984879 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:55Z","lastTransitionTime":"2025-12-06T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.088245 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.088302 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.088345 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.088368 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.088385 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:56Z","lastTransitionTime":"2025-12-06T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.191245 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.191282 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.191299 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.191356 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.191375 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:56Z","lastTransitionTime":"2025-12-06T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.295381 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.295440 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.295457 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.295484 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.295505 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:56Z","lastTransitionTime":"2025-12-06T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.398717 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.398781 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.398798 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.398823 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.398839 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:56Z","lastTransitionTime":"2025-12-06T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.501845 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.501908 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.501926 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.501950 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.501968 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:56Z","lastTransitionTime":"2025-12-06T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.510426 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:56 crc kubenswrapper[4706]: E1206 13:56:56.510589 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.604714 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.604779 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.604805 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.604834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.604856 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:56Z","lastTransitionTime":"2025-12-06T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.708171 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.708273 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.708303 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.708380 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.708407 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:56Z","lastTransitionTime":"2025-12-06T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.811816 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.811907 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.811931 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.811963 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.811987 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:56Z","lastTransitionTime":"2025-12-06T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.915908 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.916058 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.916080 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.916104 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:56 crc kubenswrapper[4706]: I1206 13:56:56.916122 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:56Z","lastTransitionTime":"2025-12-06T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.019398 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.019482 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.019521 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.019570 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.019593 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:57Z","lastTransitionTime":"2025-12-06T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.122849 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.123261 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.123540 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.123769 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.123960 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:57Z","lastTransitionTime":"2025-12-06T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.218366 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.218447 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.218466 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.218492 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.218509 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:57Z","lastTransitionTime":"2025-12-06T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:57 crc kubenswrapper[4706]: E1206 13:56:57.239226 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.244126 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.244384 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.244554 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.244712 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.244842 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:57Z","lastTransitionTime":"2025-12-06T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:57 crc kubenswrapper[4706]: E1206 13:56:57.263832 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.269534 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.269594 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.269612 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.269636 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.269654 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:57Z","lastTransitionTime":"2025-12-06T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:57 crc kubenswrapper[4706]: E1206 13:56:57.289737 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.295010 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.295082 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.295115 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.295141 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.295158 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:57Z","lastTransitionTime":"2025-12-06T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:57 crc kubenswrapper[4706]: E1206 13:56:57.314359 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.318986 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.319079 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.319100 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.319123 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.319140 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:57Z","lastTransitionTime":"2025-12-06T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:57 crc kubenswrapper[4706]: E1206 13:56:57.338175 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: E1206 13:56:57.338518 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.340684 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.340745 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.340763 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.340793 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.340812 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:57Z","lastTransitionTime":"2025-12-06T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.444155 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.444192 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.444202 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.444215 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.444226 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:57Z","lastTransitionTime":"2025-12-06T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.510236 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:57 crc kubenswrapper[4706]: E1206 13:56:57.510380 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.510441 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.510242 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:57 crc kubenswrapper[4706]: E1206 13:56:57.512442 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:57 crc kubenswrapper[4706]: E1206 13:56:57.512604 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.527877 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873f40ed-e71b-4cf8-9322-b5ddda4ec8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://631b8b4aeac64fbdd48083db4545f8df35fa2e62abb03fcfc0d78ee39216be3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0730aa973ef51ef8c4d240fde0ef6135c76ac40b576a1afa846e8b3772c16120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c44a305126e75b14cf63cc84ae01d0d04203db32abb5f4434f1339a8e773fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.547947 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.548694 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.548741 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.548752 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.548776 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.548788 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:57Z","lastTransitionTime":"2025-12-06T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.563649 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.583405 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.605098 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.620781 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5443fc82-2d54-43df-88da-043c2eb77238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhkhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.639668 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.650333 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.650356 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.650366 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.650380 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.650389 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:57Z","lastTransitionTime":"2025-12-06T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.659898 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.670369 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.684026 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b18177422ea481c7f4a00b417f08daac333e453eafde2445d257280559d50ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://272817fdf8aaa69e4b8cfa577e833c41292c789adbfaead6c4eea56351619dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.696964 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.711284 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.732690 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.747202 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.753083 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.753128 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.753142 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.753163 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.753173 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:57Z","lastTransitionTime":"2025-12-06T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.766976 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.784846 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.799295 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.826791 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://617c28d520666d053c72f94dcdb98151b0b01ce3df981cbe52903edee196d6cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"message\\\":\\\"ory.go:160\\\\nI1206 13:56:29.689017 6170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:29.689042 6170 factory.go:656] Stopping watch factory\\\\nI1206 13:56:29.689047 6170 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 13:56:29.689065 6170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:29.689075 6170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 13:56:29.689189 6170 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689457 6170 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689561 6170 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689596 6170 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:29.689629 6170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:29.689690 6170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"\\\\nI1206 13:56:47.623888 6367 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:47.624014 6367 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:47.624046 6367 factory.go:656] Stopping watch factory\\\\nI1206 13:56:47.624072 6367 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:47.624143 6367 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:47.624422 6367 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624704 6367 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624870 6367 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624895 6367 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624934 6367 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:57Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.856077 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.856161 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.856170 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.856188 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.856199 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:57Z","lastTransitionTime":"2025-12-06T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.958850 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.958921 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.958938 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.958962 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:57 crc kubenswrapper[4706]: I1206 13:56:57.958981 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:57Z","lastTransitionTime":"2025-12-06T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.062254 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.062304 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.062356 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.062401 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.062423 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:58Z","lastTransitionTime":"2025-12-06T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.165403 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.165464 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.165481 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.165504 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.165521 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:58Z","lastTransitionTime":"2025-12-06T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.268641 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.268724 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.268747 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.268783 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.268807 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:58Z","lastTransitionTime":"2025-12-06T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.371601 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.371663 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.371682 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.371706 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.371723 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:58Z","lastTransitionTime":"2025-12-06T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.474684 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.474731 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.474747 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.474770 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.474786 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:58Z","lastTransitionTime":"2025-12-06T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.511016 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:56:58 crc kubenswrapper[4706]: E1206 13:56:58.511164 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.577848 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.577920 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.577944 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.577972 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.577994 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:58Z","lastTransitionTime":"2025-12-06T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.680683 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.680764 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.680787 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.680822 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.680844 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:58Z","lastTransitionTime":"2025-12-06T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.784474 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.784785 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.784809 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.784880 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.784901 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:58Z","lastTransitionTime":"2025-12-06T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.887965 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.888294 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.888344 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.888369 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.888386 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:58Z","lastTransitionTime":"2025-12-06T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.991276 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.991363 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.991382 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.991407 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:58 crc kubenswrapper[4706]: I1206 13:56:58.991429 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:58Z","lastTransitionTime":"2025-12-06T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.094882 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.094934 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.094955 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.094979 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.094996 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:59Z","lastTransitionTime":"2025-12-06T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.197415 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.197474 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.197497 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.197524 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.197544 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:59Z","lastTransitionTime":"2025-12-06T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.300620 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.300690 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.300707 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.300732 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.300749 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:59Z","lastTransitionTime":"2025-12-06T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.403413 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.403477 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.403494 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.403520 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.403536 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:59Z","lastTransitionTime":"2025-12-06T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.506336 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.506385 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.506397 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.506415 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.506426 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:59Z","lastTransitionTime":"2025-12-06T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.510621 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.510638 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.510647 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:56:59 crc kubenswrapper[4706]: E1206 13:56:59.510725 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:56:59 crc kubenswrapper[4706]: E1206 13:56:59.510837 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:56:59 crc kubenswrapper[4706]: E1206 13:56:59.511158 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.511346 4706 scope.go:117] "RemoveContainer" containerID="dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1" Dec 06 13:56:59 crc kubenswrapper[4706]: E1206 13:56:59.511488 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-q6bgc_openshift-ovn-kubernetes(c6db0d03-fbc0-4b3b-b89d-282767c274be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.537483 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.556356 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.591581 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.609102 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.609488 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.609705 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.609904 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.610253 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:59Z","lastTransitionTime":"2025-12-06T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.613495 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.634769 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.650732 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.663926 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.689233 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"\\\\nI1206 13:56:47.623888 6367 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:47.624014 6367 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:47.624046 6367 factory.go:656] Stopping watch factory\\\\nI1206 13:56:47.624072 6367 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:47.624143 6367 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:47.624422 6367 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624704 6367 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624870 6367 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624895 6367 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624934 6367 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-q6bgc_openshift-ovn-kubernetes(c6db0d03-fbc0-4b3b-b89d-282767c274be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.699300 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873f40ed-e71b-4cf8-9322-b5ddda4ec8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://631b8b4aeac64fbdd48083db4545f8df35fa2e62abb03fcfc0d78ee39216be3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0730aa973ef51ef8c4d240fde0ef6135c76ac40b576a1afa846e8b3772c16120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c44a305126e75b14cf63cc84ae01d0d04203db32abb5f4434f1339a8e773fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.712905 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.712926 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.712934 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.712947 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.712955 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:59Z","lastTransitionTime":"2025-12-06T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.716997 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.730884 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.748274 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.766904 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.783608 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5443fc82-2d54-43df-88da-043c2eb77238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhkhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.801863 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.818597 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.819048 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.819201 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.819439 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.819615 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:59Z","lastTransitionTime":"2025-12-06T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.823738 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.837427 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.851148 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b18177422ea481c7f4a00b417f08daac333e453eafde2445d257280559d50ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://272817fdf8aaa69e4b8cfa577e833c41292c789adbfaead6c4eea56351619dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:56:59Z is after 2025-08-24T17:21:41Z" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.922366 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.922425 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.922439 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.922455 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:56:59 crc kubenswrapper[4706]: I1206 13:56:59.922467 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:56:59Z","lastTransitionTime":"2025-12-06T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.025391 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.025435 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.025452 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.025472 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.025491 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:00Z","lastTransitionTime":"2025-12-06T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.128605 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.128658 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.128677 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.128701 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.128718 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:00Z","lastTransitionTime":"2025-12-06T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.232536 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.232584 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.232600 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.232622 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.232639 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:00Z","lastTransitionTime":"2025-12-06T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.335714 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.335765 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.335781 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.335803 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.335821 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:00Z","lastTransitionTime":"2025-12-06T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.441708 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.441770 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.441793 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.441824 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.441846 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:00Z","lastTransitionTime":"2025-12-06T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.510892 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:00 crc kubenswrapper[4706]: E1206 13:57:00.511065 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.545494 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.545553 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.545569 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.545592 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.545608 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:00Z","lastTransitionTime":"2025-12-06T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.648413 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.648488 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.648504 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.648527 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.648540 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:00Z","lastTransitionTime":"2025-12-06T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.751914 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.752015 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.752068 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.752093 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.752110 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:00Z","lastTransitionTime":"2025-12-06T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.856061 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.856110 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.856125 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.856146 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.856161 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:00Z","lastTransitionTime":"2025-12-06T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.959436 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.959493 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.959514 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.959545 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:00 crc kubenswrapper[4706]: I1206 13:57:00.959567 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:00Z","lastTransitionTime":"2025-12-06T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.062781 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.062827 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.062844 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.062869 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.062886 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:01Z","lastTransitionTime":"2025-12-06T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.166183 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.166240 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.166256 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.166278 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.166294 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:01Z","lastTransitionTime":"2025-12-06T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.268692 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.268732 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.268743 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.268759 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.268770 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:01Z","lastTransitionTime":"2025-12-06T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.371591 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.371623 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.371636 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.371675 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.371685 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:01Z","lastTransitionTime":"2025-12-06T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.473957 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.474208 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.474335 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.474468 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.474585 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:01Z","lastTransitionTime":"2025-12-06T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.512259 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.512602 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.512847 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:01 crc kubenswrapper[4706]: E1206 13:57:01.513909 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:01 crc kubenswrapper[4706]: E1206 13:57:01.513342 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:01 crc kubenswrapper[4706]: E1206 13:57:01.514712 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.577631 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.577935 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.578069 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.578213 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.578370 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:01Z","lastTransitionTime":"2025-12-06T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.687924 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.688205 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.688325 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.688429 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.688570 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:01Z","lastTransitionTime":"2025-12-06T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.792253 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.792350 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.792375 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.792398 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.792416 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:01Z","lastTransitionTime":"2025-12-06T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.895869 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.896124 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.896335 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.896531 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.896678 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:01Z","lastTransitionTime":"2025-12-06T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.999583 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.999618 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.999626 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.999643 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:01 crc kubenswrapper[4706]: I1206 13:57:01.999652 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:01Z","lastTransitionTime":"2025-12-06T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.101803 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.101839 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.101849 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.101863 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.101875 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:02Z","lastTransitionTime":"2025-12-06T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.204774 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.204808 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.204819 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.204834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.204847 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:02Z","lastTransitionTime":"2025-12-06T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.307813 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.307841 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.307851 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.307865 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.307874 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:02Z","lastTransitionTime":"2025-12-06T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.410808 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.411044 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.411161 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.411239 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.411297 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:02Z","lastTransitionTime":"2025-12-06T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.510444 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:02 crc kubenswrapper[4706]: E1206 13:57:02.510633 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.515373 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.515487 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.515561 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.515641 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.515730 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:02Z","lastTransitionTime":"2025-12-06T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.618102 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.618427 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.618510 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.618607 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.618711 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:02Z","lastTransitionTime":"2025-12-06T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.721148 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.721547 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.721735 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.721888 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.722025 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:02Z","lastTransitionTime":"2025-12-06T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.824146 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.824354 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.824375 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.824397 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.824408 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:02Z","lastTransitionTime":"2025-12-06T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.927507 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.927538 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.927547 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.927561 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:02 crc kubenswrapper[4706]: I1206 13:57:02.927570 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:02Z","lastTransitionTime":"2025-12-06T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.029990 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.030050 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.030069 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.030098 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.030116 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:03Z","lastTransitionTime":"2025-12-06T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.130040 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs\") pod \"network-metrics-daemon-hhkhk\" (UID: \"5443fc82-2d54-43df-88da-043c2eb77238\") " pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:03 crc kubenswrapper[4706]: E1206 13:57:03.130285 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 13:57:03 crc kubenswrapper[4706]: E1206 13:57:03.130392 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs podName:5443fc82-2d54-43df-88da-043c2eb77238 nodeName:}" failed. No retries permitted until 2025-12-06 13:57:35.130368488 +0000 UTC m=+97.966161375 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs") pod "network-metrics-daemon-hhkhk" (UID: "5443fc82-2d54-43df-88da-043c2eb77238") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.131987 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.132022 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.132033 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.132047 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.132057 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:03Z","lastTransitionTime":"2025-12-06T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.234414 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.234456 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.234468 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.234484 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.234493 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:03Z","lastTransitionTime":"2025-12-06T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.337561 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.337599 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.337608 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.337624 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.337636 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:03Z","lastTransitionTime":"2025-12-06T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.441123 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.441160 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.441170 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.441185 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.441194 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:03Z","lastTransitionTime":"2025-12-06T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.512590 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.512773 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.512662 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:03 crc kubenswrapper[4706]: E1206 13:57:03.513064 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:03 crc kubenswrapper[4706]: E1206 13:57:03.513157 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:03 crc kubenswrapper[4706]: E1206 13:57:03.513243 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.543430 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.543457 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.543467 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.543484 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.543496 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:03Z","lastTransitionTime":"2025-12-06T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.645910 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.646499 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.646510 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.646531 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.646544 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:03Z","lastTransitionTime":"2025-12-06T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.750660 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.750711 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.750732 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.750750 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.750764 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:03Z","lastTransitionTime":"2025-12-06T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.852996 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.853047 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.853058 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.853073 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.853086 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:03Z","lastTransitionTime":"2025-12-06T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.955243 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.955535 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.955616 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.955747 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:03 crc kubenswrapper[4706]: I1206 13:57:03.955826 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:03Z","lastTransitionTime":"2025-12-06T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.058270 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.058375 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.058398 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.058431 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.058454 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:04Z","lastTransitionTime":"2025-12-06T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.160849 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.161365 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.161535 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.161693 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.161828 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:04Z","lastTransitionTime":"2025-12-06T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.264838 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.264898 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.264913 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.264938 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.264954 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:04Z","lastTransitionTime":"2025-12-06T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.366529 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.366595 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.366607 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.366627 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.366638 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:04Z","lastTransitionTime":"2025-12-06T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.470960 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.470990 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.470999 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.471012 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.471021 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:04Z","lastTransitionTime":"2025-12-06T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.510730 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:04 crc kubenswrapper[4706]: E1206 13:57:04.510871 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.573198 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.573264 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.573286 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.573355 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.573380 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:04Z","lastTransitionTime":"2025-12-06T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.675127 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.675262 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.675350 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.675423 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.675491 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:04Z","lastTransitionTime":"2025-12-06T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.777233 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.777271 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.777283 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.777297 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.777326 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:04Z","lastTransitionTime":"2025-12-06T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.879549 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.879582 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.879590 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.879602 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.879611 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:04Z","lastTransitionTime":"2025-12-06T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.981574 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.981597 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.981605 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.981617 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.981627 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:04Z","lastTransitionTime":"2025-12-06T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.983771 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xqlx6_d3fdd6dc-f817-486c-b0ee-18a6f2185559/kube-multus/0.log" Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.983922 4706 generic.go:334] "Generic (PLEG): container finished" podID="d3fdd6dc-f817-486c-b0ee-18a6f2185559" containerID="db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0" exitCode=1 Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.984003 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xqlx6" event={"ID":"d3fdd6dc-f817-486c-b0ee-18a6f2185559","Type":"ContainerDied","Data":"db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0"} Dec 06 13:57:04 crc kubenswrapper[4706]: I1206 13:57:04.984496 4706 scope.go:117] "RemoveContainer" containerID="db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.002788 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.017104 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:57:04Z\\\",\\\"message\\\":\\\"2025-12-06T13:56:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_66458ff7-3368-4e10-bee4-f5b774ad6014\\\\n2025-12-06T13:56:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_66458ff7-3368-4e10-bee4-f5b774ad6014 to /host/opt/cni/bin/\\\\n2025-12-06T13:56:19Z [verbose] multus-daemon started\\\\n2025-12-06T13:56:19Z [verbose] Readiness Indicator file check\\\\n2025-12-06T13:57:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.029630 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5443fc82-2d54-43df-88da-043c2eb77238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhkhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.048286 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873f40ed-e71b-4cf8-9322-b5ddda4ec8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://631b8b4aeac64fbdd48083db4545f8df35fa2e62abb03fcfc0d78ee39216be3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0730aa973ef51ef8c4d240fde0ef6135c76ac40b576a1afa846e8b3772c16120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c44a305126e75b14cf63cc84ae01d0d04203db32abb5f4434f1339a8e773fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.065824 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.076986 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.084959 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.084988 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.085002 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.085023 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.085041 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:05Z","lastTransitionTime":"2025-12-06T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.086417 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.097749 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b18177422ea481c7f4a00b417f08daac333e453eafde2445d257280559d50ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://272817fdf8aaa69e4b8cfa577e833c41292c789adbfaead6c4eea56351619dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.110496 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.122719 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.136788 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.149937 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.161856 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.173068 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.187537 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.187583 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.187599 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.187621 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.187640 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:05Z","lastTransitionTime":"2025-12-06T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.192994 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"\\\\nI1206 13:56:47.623888 6367 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:47.624014 6367 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:47.624046 6367 factory.go:656] Stopping watch factory\\\\nI1206 13:56:47.624072 6367 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:47.624143 6367 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:47.624422 6367 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624704 6367 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624870 6367 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624895 6367 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624934 6367 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-q6bgc_openshift-ovn-kubernetes(c6db0d03-fbc0-4b3b-b89d-282767c274be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.206331 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.228722 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.242686 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:05Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.290109 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.290156 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.290172 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.290195 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.290211 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:05Z","lastTransitionTime":"2025-12-06T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.392427 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.392465 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.392476 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.392490 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.392501 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:05Z","lastTransitionTime":"2025-12-06T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.495041 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.495134 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.495188 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.495218 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.495239 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:05Z","lastTransitionTime":"2025-12-06T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.511688 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.511707 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.511797 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:05 crc kubenswrapper[4706]: E1206 13:57:05.511820 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:05 crc kubenswrapper[4706]: E1206 13:57:05.511927 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:05 crc kubenswrapper[4706]: E1206 13:57:05.512050 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.598045 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.598072 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.598080 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.598093 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.598103 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:05Z","lastTransitionTime":"2025-12-06T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.699841 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.699885 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.699897 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.699913 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.699925 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:05Z","lastTransitionTime":"2025-12-06T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.802354 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.802410 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.802427 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.802453 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.802471 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:05Z","lastTransitionTime":"2025-12-06T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.905086 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.905147 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.905164 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.905187 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.905203 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:05Z","lastTransitionTime":"2025-12-06T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.989473 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xqlx6_d3fdd6dc-f817-486c-b0ee-18a6f2185559/kube-multus/0.log" Dec 06 13:57:05 crc kubenswrapper[4706]: I1206 13:57:05.989570 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xqlx6" event={"ID":"d3fdd6dc-f817-486c-b0ee-18a6f2185559","Type":"ContainerStarted","Data":"fcaf233bcb76d165e238dfbafde4bcbfa99d5a083980078c23ea17e2ab19d5af"} Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.008232 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.008302 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.008350 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.008587 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.008632 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:06Z","lastTransitionTime":"2025-12-06T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.010536 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.032871 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.052010 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.065055 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.080886 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.094839 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.113141 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.113170 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.113178 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.113192 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.113201 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:06Z","lastTransitionTime":"2025-12-06T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.114767 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"\\\\nI1206 13:56:47.623888 6367 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:47.624014 6367 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:47.624046 6367 factory.go:656] Stopping watch factory\\\\nI1206 13:56:47.624072 6367 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:47.624143 6367 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:47.624422 6367 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624704 6367 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624870 6367 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624895 6367 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624934 6367 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-q6bgc_openshift-ovn-kubernetes(c6db0d03-fbc0-4b3b-b89d-282767c274be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.131651 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873f40ed-e71b-4cf8-9322-b5ddda4ec8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://631b8b4aeac64fbdd48083db4545f8df35fa2e62abb03fcfc0d78ee39216be3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0730aa973ef51ef8c4d240fde0ef6135c76ac40b576a1afa846e8b3772c16120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c44a305126e75b14cf63cc84ae01d0d04203db32abb5f4434f1339a8e773fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.148917 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.161154 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.178168 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.191667 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcaf233bcb76d165e238dfbafde4bcbfa99d5a083980078c23ea17e2ab19d5af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:57:04Z\\\",\\\"message\\\":\\\"2025-12-06T13:56:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_66458ff7-3368-4e10-bee4-f5b774ad6014\\\\n2025-12-06T13:56:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_66458ff7-3368-4e10-bee4-f5b774ad6014 to /host/opt/cni/bin/\\\\n2025-12-06T13:56:19Z [verbose] multus-daemon started\\\\n2025-12-06T13:56:19Z [verbose] Readiness Indicator file check\\\\n2025-12-06T13:57:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.208020 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5443fc82-2d54-43df-88da-043c2eb77238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhkhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.215644 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.215681 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.215691 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.215709 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.215721 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:06Z","lastTransitionTime":"2025-12-06T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.229630 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.248048 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.259726 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.273020 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b18177422ea481c7f4a00b417f08daac333e453eafde2445d257280559d50ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://272817fdf8aaa69e4b8cfa577e833c41292c789adbfaead6c4eea56351619dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.285203 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:06Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.318189 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.318245 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.318260 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.318280 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.318297 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:06Z","lastTransitionTime":"2025-12-06T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.420737 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.420796 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.420818 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.420846 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.420869 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:06Z","lastTransitionTime":"2025-12-06T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.510883 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:06 crc kubenswrapper[4706]: E1206 13:57:06.511021 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.523474 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.523530 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.523552 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.523579 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.523601 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:06Z","lastTransitionTime":"2025-12-06T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.625844 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.625876 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.625886 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.625899 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.625910 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:06Z","lastTransitionTime":"2025-12-06T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.729150 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.729209 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.729222 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.729239 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.729253 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:06Z","lastTransitionTime":"2025-12-06T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.832268 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.832382 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.832408 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.832436 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.832459 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:06Z","lastTransitionTime":"2025-12-06T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.935048 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.935105 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.935118 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.935137 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:06 crc kubenswrapper[4706]: I1206 13:57:06.935153 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:06Z","lastTransitionTime":"2025-12-06T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.037581 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.037625 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.037639 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.037656 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.037666 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:07Z","lastTransitionTime":"2025-12-06T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.139333 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.139393 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.139411 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.139457 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.139476 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:07Z","lastTransitionTime":"2025-12-06T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.241847 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.241898 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.241909 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.241924 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.241935 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:07Z","lastTransitionTime":"2025-12-06T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.343473 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.343504 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.343516 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.343530 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.343542 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:07Z","lastTransitionTime":"2025-12-06T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.445291 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.445351 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.445362 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.445376 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.445385 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:07Z","lastTransitionTime":"2025-12-06T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.510494 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.510525 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:07 crc kubenswrapper[4706]: E1206 13:57:07.510704 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.510723 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:07 crc kubenswrapper[4706]: E1206 13:57:07.510794 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:07 crc kubenswrapper[4706]: E1206 13:57:07.510887 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.522301 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873f40ed-e71b-4cf8-9322-b5ddda4ec8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://631b8b4aeac64fbdd48083db4545f8df35fa2e62abb03fcfc0d78ee39216be3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0730aa973ef51ef8c4d240fde0ef6135c76ac40b576a1afa846e8b3772c16120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c44a305126e75b14cf63cc84ae01d0d04203db32abb5f4434f1339a8e773fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.533116 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.544839 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.547972 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.548012 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.548021 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.548035 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.548044 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:07Z","lastTransitionTime":"2025-12-06T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.560694 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.580226 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcaf233bcb76d165e238dfbafde4bcbfa99d5a083980078c23ea17e2ab19d5af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:57:04Z\\\",\\\"message\\\":\\\"2025-12-06T13:56:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_66458ff7-3368-4e10-bee4-f5b774ad6014\\\\n2025-12-06T13:56:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_66458ff7-3368-4e10-bee4-f5b774ad6014 to /host/opt/cni/bin/\\\\n2025-12-06T13:56:19Z [verbose] multus-daemon started\\\\n2025-12-06T13:56:19Z [verbose] Readiness Indicator file check\\\\n2025-12-06T13:57:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.593790 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5443fc82-2d54-43df-88da-043c2eb77238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhkhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.609437 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.630366 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.642469 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.650902 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.650939 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.650950 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.650967 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.650981 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:07Z","lastTransitionTime":"2025-12-06T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.654811 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b18177422ea481c7f4a00b417f08daac333e453eafde2445d257280559d50ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://272817fdf8aaa69e4b8cfa577e833c41292c789adbfaead6c4eea56351619dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.668254 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.685542 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.700379 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.700425 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.700435 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.700450 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.700460 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:07Z","lastTransitionTime":"2025-12-06T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.706476 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: E1206 13:57:07.718423 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.721576 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.721912 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.721995 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.722021 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.722056 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.722081 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:07Z","lastTransitionTime":"2025-12-06T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.733283 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: E1206 13:57:07.737867 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.741131 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.741236 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.741294 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.741389 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.741450 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:07Z","lastTransitionTime":"2025-12-06T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.746481 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: E1206 13:57:07.752686 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.755625 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.755670 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.755684 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.755700 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.755712 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:07Z","lastTransitionTime":"2025-12-06T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.756104 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.773754 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"\\\\nI1206 13:56:47.623888 6367 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:47.624014 6367 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:47.624046 6367 factory.go:656] Stopping watch factory\\\\nI1206 13:56:47.624072 6367 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:47.624143 6367 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:47.624422 6367 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624704 6367 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624870 6367 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624895 6367 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624934 6367 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-q6bgc_openshift-ovn-kubernetes(c6db0d03-fbc0-4b3b-b89d-282767c274be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: E1206 13:57:07.774826 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.778898 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.778935 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.778948 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.778965 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.778978 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:07Z","lastTransitionTime":"2025-12-06T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:07 crc kubenswrapper[4706]: E1206 13:57:07.797075 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:07Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:07 crc kubenswrapper[4706]: E1206 13:57:07.797773 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.799741 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.799832 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.799908 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.799967 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.800030 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:07Z","lastTransitionTime":"2025-12-06T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.902504 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.902571 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.902587 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.902613 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:07 crc kubenswrapper[4706]: I1206 13:57:07.902630 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:07Z","lastTransitionTime":"2025-12-06T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.005154 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.005217 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.005235 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.005261 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.005283 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:08Z","lastTransitionTime":"2025-12-06T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.107489 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.107535 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.107551 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.107574 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.107591 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:08Z","lastTransitionTime":"2025-12-06T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.210102 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.210152 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.210164 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.210181 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.210193 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:08Z","lastTransitionTime":"2025-12-06T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.312362 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.312486 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.312549 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.312639 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.312724 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:08Z","lastTransitionTime":"2025-12-06T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.415067 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.415106 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.415114 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.415129 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.415138 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:08Z","lastTransitionTime":"2025-12-06T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.510294 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:08 crc kubenswrapper[4706]: E1206 13:57:08.510482 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.516856 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.516890 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.516904 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.516918 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.516931 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:08Z","lastTransitionTime":"2025-12-06T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.618508 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.618825 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.618952 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.619080 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.619183 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:08Z","lastTransitionTime":"2025-12-06T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.722255 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.722631 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.722819 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.723055 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.723264 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:08Z","lastTransitionTime":"2025-12-06T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.825969 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.826013 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.826021 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.826035 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.826044 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:08Z","lastTransitionTime":"2025-12-06T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.928206 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.928250 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.928261 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.928277 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:08 crc kubenswrapper[4706]: I1206 13:57:08.928290 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:08Z","lastTransitionTime":"2025-12-06T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.031282 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.031365 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.031377 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.031394 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.031430 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:09Z","lastTransitionTime":"2025-12-06T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.133997 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.134043 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.134059 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.134082 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.134099 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:09Z","lastTransitionTime":"2025-12-06T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.237020 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.237070 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.237087 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.237110 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.237126 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:09Z","lastTransitionTime":"2025-12-06T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.339615 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.339642 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.339651 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.339665 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.339674 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:09Z","lastTransitionTime":"2025-12-06T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.443119 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.443148 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.443156 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.443174 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.443184 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:09Z","lastTransitionTime":"2025-12-06T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.510081 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.510087 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.510088 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:09 crc kubenswrapper[4706]: E1206 13:57:09.510358 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:09 crc kubenswrapper[4706]: E1206 13:57:09.510443 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:09 crc kubenswrapper[4706]: E1206 13:57:09.510210 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.545694 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.545993 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.546215 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.546488 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.546689 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:09Z","lastTransitionTime":"2025-12-06T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.649882 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.650264 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.650525 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.650774 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.650969 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:09Z","lastTransitionTime":"2025-12-06T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.754365 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.754422 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.754440 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.754465 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.754488 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:09Z","lastTransitionTime":"2025-12-06T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.857581 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.857662 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.857711 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.857733 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.857749 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:09Z","lastTransitionTime":"2025-12-06T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.961302 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.961389 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.961405 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.961428 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:09 crc kubenswrapper[4706]: I1206 13:57:09.961446 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:09Z","lastTransitionTime":"2025-12-06T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.064126 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.064175 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.064192 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.064215 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.064232 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:10Z","lastTransitionTime":"2025-12-06T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.166883 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.166943 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.166960 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.166983 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.167000 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:10Z","lastTransitionTime":"2025-12-06T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.268994 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.269134 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.269156 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.269180 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.269196 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:10Z","lastTransitionTime":"2025-12-06T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.371329 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.371411 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.371437 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.371464 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.371482 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:10Z","lastTransitionTime":"2025-12-06T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.474617 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.474681 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.474703 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.474727 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.474744 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:10Z","lastTransitionTime":"2025-12-06T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.510734 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:10 crc kubenswrapper[4706]: E1206 13:57:10.510996 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.512042 4706 scope.go:117] "RemoveContainer" containerID="dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.577227 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.577289 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.577348 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.577394 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.577421 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:10Z","lastTransitionTime":"2025-12-06T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.680701 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.680931 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.680941 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.680953 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.680963 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:10Z","lastTransitionTime":"2025-12-06T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.783881 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.783965 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.783990 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.784024 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.784044 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:10Z","lastTransitionTime":"2025-12-06T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.886574 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.886609 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.886617 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.886633 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.886643 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:10Z","lastTransitionTime":"2025-12-06T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.988626 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.988696 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.988713 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.988738 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:10 crc kubenswrapper[4706]: I1206 13:57:10.988756 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:10Z","lastTransitionTime":"2025-12-06T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.006926 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovnkube-controller/2.log" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.009833 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerStarted","Data":"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17"} Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.010364 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.027917 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.046791 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"\\\\nI1206 13:56:47.623888 6367 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:47.624014 6367 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:47.624046 6367 factory.go:656] Stopping watch factory\\\\nI1206 13:56:47.624072 6367 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:47.624143 6367 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:47.624422 6367 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624704 6367 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624870 6367 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624895 6367 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624934 6367 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.071124 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.091160 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.091210 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.091223 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.091241 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.091255 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:11Z","lastTransitionTime":"2025-12-06T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.095838 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.117520 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.168591 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.190860 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.193685 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.193755 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.193778 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.193804 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.193823 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:11Z","lastTransitionTime":"2025-12-06T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.210387 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.229521 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873f40ed-e71b-4cf8-9322-b5ddda4ec8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://631b8b4aeac64fbdd48083db4545f8df35fa2e62abb03fcfc0d78ee39216be3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0730aa973ef51ef8c4d240fde0ef6135c76ac40b576a1afa846e8b3772c16120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c44a305126e75b14cf63cc84ae01d0d04203db32abb5f4434f1339a8e773fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.248230 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.268497 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.294521 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.295792 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.295823 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.295834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.295852 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.295873 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:11Z","lastTransitionTime":"2025-12-06T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.310241 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcaf233bcb76d165e238dfbafde4bcbfa99d5a083980078c23ea17e2ab19d5af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:57:04Z\\\",\\\"message\\\":\\\"2025-12-06T13:56:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_66458ff7-3368-4e10-bee4-f5b774ad6014\\\\n2025-12-06T13:56:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_66458ff7-3368-4e10-bee4-f5b774ad6014 to /host/opt/cni/bin/\\\\n2025-12-06T13:56:19Z [verbose] multus-daemon started\\\\n2025-12-06T13:56:19Z [verbose] Readiness Indicator file check\\\\n2025-12-06T13:57:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.321801 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5443fc82-2d54-43df-88da-043c2eb77238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhkhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.333398 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.352383 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.362257 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.372527 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b18177422ea481c7f4a00b417f08daac333e453eafde2445d257280559d50ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://272817fdf8aaa69e4b8cfa577e833c41292c789adbfaead6c4eea56351619dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:11Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.398799 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.398848 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.398861 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.398879 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.398891 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:11Z","lastTransitionTime":"2025-12-06T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.500991 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.501040 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.501058 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.501081 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.501095 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:11Z","lastTransitionTime":"2025-12-06T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.510671 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.510753 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:11 crc kubenswrapper[4706]: E1206 13:57:11.510823 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:11 crc kubenswrapper[4706]: E1206 13:57:11.510888 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.510946 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:11 crc kubenswrapper[4706]: E1206 13:57:11.511091 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.604204 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.604263 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.604282 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.604343 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.604364 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:11Z","lastTransitionTime":"2025-12-06T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.707915 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.707959 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.707973 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.707989 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.708000 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:11Z","lastTransitionTime":"2025-12-06T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.811200 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.811252 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.811270 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.811293 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.811352 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:11Z","lastTransitionTime":"2025-12-06T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.913725 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.913788 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.913808 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.913833 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:11 crc kubenswrapper[4706]: I1206 13:57:11.913854 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:11Z","lastTransitionTime":"2025-12-06T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.016774 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.016885 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.016907 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.016931 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.016948 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:12Z","lastTransitionTime":"2025-12-06T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.123950 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.124023 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.124058 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.124211 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.124243 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:12Z","lastTransitionTime":"2025-12-06T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.228395 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.228449 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.228465 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.228485 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.228503 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:12Z","lastTransitionTime":"2025-12-06T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.330732 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.330780 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.330801 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.330827 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.330850 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:12Z","lastTransitionTime":"2025-12-06T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.433939 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.434010 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.434027 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.434051 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.434069 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:12Z","lastTransitionTime":"2025-12-06T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.510578 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:12 crc kubenswrapper[4706]: E1206 13:57:12.510774 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.537121 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.537181 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.537198 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.537220 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.537237 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:12Z","lastTransitionTime":"2025-12-06T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.640373 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.640437 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.640460 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.640486 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.640505 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:12Z","lastTransitionTime":"2025-12-06T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.743139 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.743233 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.743289 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.743384 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.743444 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:12Z","lastTransitionTime":"2025-12-06T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.846821 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.846880 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.846896 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.846918 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.846937 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:12Z","lastTransitionTime":"2025-12-06T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.950359 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.950411 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.950428 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.950453 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:12 crc kubenswrapper[4706]: I1206 13:57:12.950472 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:12Z","lastTransitionTime":"2025-12-06T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.020295 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovnkube-controller/3.log" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.023271 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovnkube-controller/2.log" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.027413 4706 generic.go:334] "Generic (PLEG): container finished" podID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerID="aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17" exitCode=1 Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.027475 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerDied","Data":"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17"} Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.027563 4706 scope.go:117] "RemoveContainer" containerID="dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.029416 4706 scope.go:117] "RemoveContainer" containerID="aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17" Dec 06 13:57:13 crc kubenswrapper[4706]: E1206 13:57:13.029841 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-q6bgc_openshift-ovn-kubernetes(c6db0d03-fbc0-4b3b-b89d-282767c274be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.053389 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.053473 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.053498 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.053529 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.053553 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:13Z","lastTransitionTime":"2025-12-06T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.058467 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.093965 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.115664 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.135634 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.153364 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.157398 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.157450 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.157469 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.157494 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.157512 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:13Z","lastTransitionTime":"2025-12-06T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.168249 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.201198 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"\\\\nI1206 13:56:47.623888 6367 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:47.624014 6367 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:47.624046 6367 factory.go:656] Stopping watch factory\\\\nI1206 13:56:47.624072 6367 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:47.624143 6367 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:47.624422 6367 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624704 6367 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624870 6367 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624895 6367 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624934 6367 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:57:12Z\\\",\\\"message\\\":\\\"er.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 13:57:11.742291 6717 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 13:57:11.742336 6717 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 13:57:11.742351 6717 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 13:57:11.742287 6717 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 13:57:11.742370 6717 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 13:57:11.742393 6717 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 13:57:11.742395 6717 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 13:57:11.742418 6717 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 13:57:11.742444 6717 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:57:11.742519 6717 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 13:57:11.742555 6717 factory.go:656] Stopping watch factory\\\\nI1206 13:57:11.742578 6717 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:57:11.742540 6717 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 13:57:11.742709 6717 ovnkube.go:599] Stopped ovnkube\\\\nI1206 13:57:11.742786 6717 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 13:57:11.742944 6717 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.222046 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.242254 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.260858 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.260922 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.260939 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.260964 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.260981 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:13Z","lastTransitionTime":"2025-12-06T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.262016 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.284223 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.305009 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcaf233bcb76d165e238dfbafde4bcbfa99d5a083980078c23ea17e2ab19d5af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:57:04Z\\\",\\\"message\\\":\\\"2025-12-06T13:56:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_66458ff7-3368-4e10-bee4-f5b774ad6014\\\\n2025-12-06T13:56:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_66458ff7-3368-4e10-bee4-f5b774ad6014 to /host/opt/cni/bin/\\\\n2025-12-06T13:56:19Z [verbose] multus-daemon started\\\\n2025-12-06T13:56:19Z [verbose] Readiness Indicator file check\\\\n2025-12-06T13:57:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.322215 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5443fc82-2d54-43df-88da-043c2eb77238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhkhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.341097 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873f40ed-e71b-4cf8-9322-b5ddda4ec8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://631b8b4aeac64fbdd48083db4545f8df35fa2e62abb03fcfc0d78ee39216be3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0730aa973ef51ef8c4d240fde0ef6135c76ac40b576a1afa846e8b3772c16120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c44a305126e75b14cf63cc84ae01d0d04203db32abb5f4434f1339a8e773fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.361832 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.363893 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.363949 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.363971 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.364001 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.364020 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:13Z","lastTransitionTime":"2025-12-06T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.381922 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.398276 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.414271 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b18177422ea481c7f4a00b417f08daac333e453eafde2445d257280559d50ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://272817fdf8aaa69e4b8cfa577e833c41292c789adbfaead6c4eea56351619dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:13Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.467406 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.467482 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.467510 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.467540 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.467563 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:13Z","lastTransitionTime":"2025-12-06T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.510472 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.510607 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:13 crc kubenswrapper[4706]: E1206 13:57:13.510643 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.510471 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:13 crc kubenswrapper[4706]: E1206 13:57:13.510816 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:13 crc kubenswrapper[4706]: E1206 13:57:13.510977 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.571554 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.571684 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.571702 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.571730 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.571758 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:13Z","lastTransitionTime":"2025-12-06T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.674604 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.674668 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.674685 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.674711 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.674729 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:13Z","lastTransitionTime":"2025-12-06T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.777902 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.778122 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.778151 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.778179 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.778197 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:13Z","lastTransitionTime":"2025-12-06T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.881094 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.881173 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.881198 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.881230 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.881256 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:13Z","lastTransitionTime":"2025-12-06T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.984368 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.984439 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.984463 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.984491 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:13 crc kubenswrapper[4706]: I1206 13:57:13.984513 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:13Z","lastTransitionTime":"2025-12-06T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.032699 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovnkube-controller/3.log" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.087155 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.087254 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.087274 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.087381 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.087410 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:14Z","lastTransitionTime":"2025-12-06T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.189934 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.189991 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.190009 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.190032 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.190056 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:14Z","lastTransitionTime":"2025-12-06T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.292763 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.292825 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.292844 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.292868 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.292888 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:14Z","lastTransitionTime":"2025-12-06T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.395057 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.395116 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.395133 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.395201 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.395227 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:14Z","lastTransitionTime":"2025-12-06T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.498446 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.498501 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.498524 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.498552 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.498576 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:14Z","lastTransitionTime":"2025-12-06T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.510972 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:14 crc kubenswrapper[4706]: E1206 13:57:14.511125 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.602084 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.602132 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.602145 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.602163 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.602177 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:14Z","lastTransitionTime":"2025-12-06T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.704483 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.704522 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.704534 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.704550 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.704562 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:14Z","lastTransitionTime":"2025-12-06T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.808063 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.808129 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.808149 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.808178 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.808196 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:14Z","lastTransitionTime":"2025-12-06T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.911071 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.911120 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.911131 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.911148 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:14 crc kubenswrapper[4706]: I1206 13:57:14.911160 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:14Z","lastTransitionTime":"2025-12-06T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.013608 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.013666 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.013684 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.013707 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.013724 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:15Z","lastTransitionTime":"2025-12-06T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.116527 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.116602 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.116620 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.116644 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.116662 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:15Z","lastTransitionTime":"2025-12-06T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.219918 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.220014 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.220069 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.220094 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.220113 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:15Z","lastTransitionTime":"2025-12-06T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.323222 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.323275 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.323291 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.323340 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.323361 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:15Z","lastTransitionTime":"2025-12-06T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.426431 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.426495 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.426512 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.426540 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.426557 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:15Z","lastTransitionTime":"2025-12-06T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.510448 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.510511 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.510456 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:15 crc kubenswrapper[4706]: E1206 13:57:15.510691 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:15 crc kubenswrapper[4706]: E1206 13:57:15.510870 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:15 crc kubenswrapper[4706]: E1206 13:57:15.511027 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.529529 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.529586 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.529603 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.529657 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.529675 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:15Z","lastTransitionTime":"2025-12-06T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.632657 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.632981 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.633014 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.633042 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.633063 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:15Z","lastTransitionTime":"2025-12-06T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.735963 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.736375 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.736552 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.736714 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.736860 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:15Z","lastTransitionTime":"2025-12-06T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.841059 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.841133 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.841184 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.841217 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.841239 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:15Z","lastTransitionTime":"2025-12-06T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.946095 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.946204 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.946230 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.946264 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:15 crc kubenswrapper[4706]: I1206 13:57:15.946288 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:15Z","lastTransitionTime":"2025-12-06T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.049814 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.049868 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.049888 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.050092 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.050110 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:16Z","lastTransitionTime":"2025-12-06T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.153488 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.153975 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.154197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.154455 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.154669 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:16Z","lastTransitionTime":"2025-12-06T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.258074 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.258146 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.258170 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.258202 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.258224 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:16Z","lastTransitionTime":"2025-12-06T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.361850 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.361903 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.361922 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.361945 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.361961 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:16Z","lastTransitionTime":"2025-12-06T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.465689 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.465769 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.465788 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.465821 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.465846 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:16Z","lastTransitionTime":"2025-12-06T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.510911 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:16 crc kubenswrapper[4706]: E1206 13:57:16.511367 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.569274 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.569668 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.569830 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.569984 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.570130 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:16Z","lastTransitionTime":"2025-12-06T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.674058 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.674524 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.674883 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.675197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.676262 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:16Z","lastTransitionTime":"2025-12-06T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.779229 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.779303 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.779358 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.779389 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.779415 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:16Z","lastTransitionTime":"2025-12-06T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.882444 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.882518 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.882539 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.882564 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.882582 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:16Z","lastTransitionTime":"2025-12-06T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.985597 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.985960 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.986113 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.986270 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:16 crc kubenswrapper[4706]: I1206 13:57:16.986460 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:16Z","lastTransitionTime":"2025-12-06T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.088604 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.088645 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.088655 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.088671 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.088682 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:17Z","lastTransitionTime":"2025-12-06T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.192006 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.192049 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.192068 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.192091 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.192108 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:17Z","lastTransitionTime":"2025-12-06T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.295661 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.295728 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.295746 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.295772 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.295789 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:17Z","lastTransitionTime":"2025-12-06T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.398528 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.398596 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.398614 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.398642 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.398662 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:17Z","lastTransitionTime":"2025-12-06T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.501993 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.502060 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.502083 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.502114 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.502134 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:17Z","lastTransitionTime":"2025-12-06T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.510451 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.510518 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:17 crc kubenswrapper[4706]: E1206 13:57:17.510667 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:17 crc kubenswrapper[4706]: E1206 13:57:17.511053 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.511431 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:17 crc kubenswrapper[4706]: E1206 13:57:17.511617 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.526907 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-49gn6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4fab411c-d62e-4003-b184-77b45c252e9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84f16881a6cec892490e29d4484e48c4b8ba4ae4b427a68f70dbb16654f4b7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fhb9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-49gn6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.559651 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6db0d03-fbc0-4b3b-b89d-282767c274be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea4139761e3eeab705d799becf02edcb4784f2dc09e81ebd544bcfc8af21bd1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:56:47Z\\\",\\\"message\\\":\\\"\\\\nI1206 13:56:47.623888 6367 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:56:47.624014 6367 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 13:56:47.624046 6367 factory.go:656] Stopping watch factory\\\\nI1206 13:56:47.624072 6367 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:56:47.624143 6367 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 13:56:47.624422 6367 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624704 6367 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624870 6367 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624895 6367 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 13:56:47.624934 6367 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:57:12Z\\\",\\\"message\\\":\\\"er.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 13:57:11.742291 6717 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 13:57:11.742336 6717 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 13:57:11.742351 6717 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 13:57:11.742287 6717 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 13:57:11.742370 6717 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 13:57:11.742393 6717 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 13:57:11.742395 6717 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 13:57:11.742418 6717 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 13:57:11.742444 6717 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 13:57:11.742519 6717 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 13:57:11.742555 6717 factory.go:656] Stopping watch factory\\\\nI1206 13:57:11.742578 6717 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 13:57:11.742540 6717 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 13:57:11.742709 6717 ovnkube.go:599] Stopped ovnkube\\\\nI1206 13:57:11.742786 6717 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 13:57:11.742944 6717 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnr6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q6bgc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.579653 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638fa176-66fa-41f4-b7f9-da645d549014\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2ea185d7efa024f60a6515d88583fda643e9a8ca630ea6c743a42714afd391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e7b8e0b5554ce689621fcd3c6116ff99390046cc8b435664a065e098585e318\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce32d79ec8c205f2e1d628c6b5faf1d6d5e5eb37bdc6472b90a21ac87200c75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.605152 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.605209 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.605257 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.605280 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.605296 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:17Z","lastTransitionTime":"2025-12-06T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.613571 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b83cb904-d414-4d60-a6e1-74413b8d7004\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23346acd1105bd422a55e47d1c016edcfaa2f040b3902c8c8520aa4f1292622b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df790ccee184ed3c146c78ad915009a9e9f08d23a5d4f09346f0bc622237261d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281d1c5eafa9828dee87f6be14055b35b402551eed0462228b9b95f1562e35d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77ce4855b2f5c1e1cfc49cd3f356766201cb9c9813b7ab3a7f8fbe2cb09da913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ae053c358c791cec21e51fd34327f480c82128d4943343592ec62715068753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54adc348444828190f149348ed549602efba37f595750a5126b2276680b843fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac6be285070fc575d7976d954d38d6fac5ab200cf8027b2c91df3919d8d5dc4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87ea5e46fd4ae4fda72346a6ab41493348dce256d50b06a8e8c8358eb044848b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.637536 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e865c24c-0cb5-40dc-8b60-d30a6e80780d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T13:56:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 13:56:09.964364 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 13:56:09.966087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2185334392/tls.crt::/tmp/serving-cert-2185334392/tls.key\\\\\\\"\\\\nI1206 13:56:15.574493 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 13:56:15.577948 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 13:56:15.577986 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 13:56:15.578030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 13:56:15.578066 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 13:56:15.590520 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 13:56:15.590680 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590746 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 13:56:15.590812 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 13:56:15.590874 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 13:56:15.590935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 13:56:15.590993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 13:56:15.591413 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 13:56:15.599378 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.656360 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.673257 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"904f5a16-0c23-4a2a-a945-59b1142cb4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a79105af018756c296bf563a86415e43973b37dffde2aaf26764a3bff538eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsw72\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-67v56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.690248 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5443fc82-2d54-43df-88da-043c2eb77238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swfnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhkhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.707482 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873f40ed-e71b-4cf8-9322-b5ddda4ec8fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://631b8b4aeac64fbdd48083db4545f8df35fa2e62abb03fcfc0d78ee39216be3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0730aa973ef51ef8c4d240fde0ef6135c76ac40b576a1afa846e8b3772c16120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c44a305126e75b14cf63cc84ae01d0d04203db32abb5f4434f1339a8e773fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:55:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034f7daa99399750636fb8b38d6e9f5df98250c8c1775db1c58556658b99f607\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:55:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:55:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:55:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.708864 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.708928 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.708944 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.708970 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.708988 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:17Z","lastTransitionTime":"2025-12-06T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.727060 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.743934 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e735cde0dc05731b4fc15eaaf9864fa8522ff2d034cd1db80b97dc7ffe12b353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.764415 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://accedace94059eb4029a4ae80d0e13632f85e798ae2a08f497ddc02a10ef8a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35a847bceb994403e2a3e93d15e6bcefc8fbf0e9f2a69e984c267ed80d217e65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.786548 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xqlx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3fdd6dc-f817-486c-b0ee-18a6f2185559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcaf233bcb76d165e238dfbafde4bcbfa99d5a083980078c23ea17e2ab19d5af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T13:57:04Z\\\",\\\"message\\\":\\\"2025-12-06T13:56:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_66458ff7-3368-4e10-bee4-f5b774ad6014\\\\n2025-12-06T13:56:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_66458ff7-3368-4e10-bee4-f5b774ad6014 to /host/opt/cni/bin/\\\\n2025-12-06T13:56:19Z [verbose] multus-daemon started\\\\n2025-12-06T13:56:19Z [verbose] Readiness Indicator file check\\\\n2025-12-06T13:57:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-thxfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xqlx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.808931 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://283aa1808fa4ac4051682dd84479eeea8ff976370a281860bd85f10fd27dca9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.812830 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.812879 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.812896 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.812921 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.812939 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:17Z","lastTransitionTime":"2025-12-06T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.833222 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80e25f9c-2df9-42c2-b82a-ecc877b7fd32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95c6630d0af28562df5a4f1098d6904b30af9e15650c5ca4fa82254071c492fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf21f18d9d9828e95c16854cc7e65237e3b43b7010a4c80e4837e137163e1f74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb051754939090e8f685461b560e569d3f18a8d4af97d9a6e80cc237cf0045e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1b45b5a789a1c1081f108b22fd3f6c36d9190184bc046d6d48ea819e61ab4a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f81d22a53ff2ed1c6253460282f75311853f9f285401fae894cfab403b781d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f847d6ff0d10be22e6b65df8c5ff104abcea3d574f1a42c58cddafafb0813f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c50f5bf79cff712026c8b77d07f30630fb9a50f845aaffc0971024fcc30bbb16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T13:56:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T13:56:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7dm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9rrsc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.849766 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2cc78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa340a3c-fe06-4af7-ab8d-77e223027038\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a704671e8e0c7dbafec4e8d9584821ddf2538c194fdd577f3c71d83fc70a17a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8bz6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2cc78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.867557 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a34577b8-1ea8-4af6-8585-0d9710dd30ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b18177422ea481c7f4a00b417f08daac333e453eafde2445d257280559d50ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://272817fdf8aaa69e4b8cfa577e833c41292c789adbfaead6c4eea56351619dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T13:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g54pg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T13:56:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-dzzz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.886569 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T13:56:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.915782 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.915825 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.915841 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.915865 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.915882 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:17Z","lastTransitionTime":"2025-12-06T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.964661 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.964831 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.964975 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.965115 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.965241 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:17Z","lastTransitionTime":"2025-12-06T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:17 crc kubenswrapper[4706]: E1206 13:57:17.985415 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:17Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.990643 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.990834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.990974 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.991163 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:17 crc kubenswrapper[4706]: I1206 13:57:17.991291 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:17Z","lastTransitionTime":"2025-12-06T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:18 crc kubenswrapper[4706]: E1206 13:57:18.010388 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.015012 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.015063 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.015077 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.015095 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.015108 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:18Z","lastTransitionTime":"2025-12-06T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:18 crc kubenswrapper[4706]: E1206 13:57:18.033633 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.038750 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.038792 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.038804 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.038822 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.038833 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:18Z","lastTransitionTime":"2025-12-06T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:18 crc kubenswrapper[4706]: E1206 13:57:18.060713 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.065815 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.065856 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.065877 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.065901 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.065918 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:18Z","lastTransitionTime":"2025-12-06T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:18 crc kubenswrapper[4706]: E1206 13:57:18.085042 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T13:57:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d0ebee1-8e8f-40f3-be2d-d81f84254664\\\",\\\"systemUUID\\\":\\\"0e9ac240-47ff-4359-8abf-1e7458f9a542\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T13:57:18Z is after 2025-08-24T17:21:41Z" Dec 06 13:57:18 crc kubenswrapper[4706]: E1206 13:57:18.085198 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.087176 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.087214 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.087225 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.087241 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.087254 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:18Z","lastTransitionTime":"2025-12-06T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.190089 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.190151 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.190172 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.190201 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.190221 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:18Z","lastTransitionTime":"2025-12-06T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.293737 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.293790 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.293808 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.293829 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.293867 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:18Z","lastTransitionTime":"2025-12-06T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.397062 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.397128 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.397147 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.397173 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.397191 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:18Z","lastTransitionTime":"2025-12-06T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.499918 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.499959 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.499969 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.499983 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.499993 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:18Z","lastTransitionTime":"2025-12-06T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.510324 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:18 crc kubenswrapper[4706]: E1206 13:57:18.510451 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.603382 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.603460 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.603486 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.603515 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.603539 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:18Z","lastTransitionTime":"2025-12-06T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.708535 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.708590 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.708609 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.708635 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.708654 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:18Z","lastTransitionTime":"2025-12-06T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.815653 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.815720 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.815737 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.815762 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.815788 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:18Z","lastTransitionTime":"2025-12-06T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.922617 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.922652 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.922662 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.922677 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:18 crc kubenswrapper[4706]: I1206 13:57:18.922688 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:18Z","lastTransitionTime":"2025-12-06T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.025620 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.025726 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.025746 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.025810 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.025828 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:19Z","lastTransitionTime":"2025-12-06T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.129607 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.129659 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.129677 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.129699 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.129715 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:19Z","lastTransitionTime":"2025-12-06T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.233279 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.233386 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.233404 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.233939 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.233987 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:19Z","lastTransitionTime":"2025-12-06T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.337910 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.338034 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.338054 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.338078 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.338095 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:19Z","lastTransitionTime":"2025-12-06T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.442483 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.442560 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.442579 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.443008 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.443034 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:19Z","lastTransitionTime":"2025-12-06T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.510764 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:19 crc kubenswrapper[4706]: E1206 13:57:19.510899 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.510780 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.511161 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:19 crc kubenswrapper[4706]: E1206 13:57:19.511189 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:19 crc kubenswrapper[4706]: E1206 13:57:19.511609 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.546757 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.546806 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.546825 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.546847 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.546864 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:19Z","lastTransitionTime":"2025-12-06T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.649627 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.649726 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.649748 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.649783 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.649805 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:19Z","lastTransitionTime":"2025-12-06T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.753550 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.753948 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.754102 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.754259 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.754444 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:19Z","lastTransitionTime":"2025-12-06T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.858559 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.858647 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.858670 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.858698 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.858717 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:19Z","lastTransitionTime":"2025-12-06T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.961969 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.962057 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.962077 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.962112 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:19 crc kubenswrapper[4706]: I1206 13:57:19.962150 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:19Z","lastTransitionTime":"2025-12-06T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.064605 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.064661 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.064678 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.064704 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.064725 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:20Z","lastTransitionTime":"2025-12-06T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.167818 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.167899 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.167919 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.167951 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.167971 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:20Z","lastTransitionTime":"2025-12-06T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.270721 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.270788 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.270809 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.270834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.270857 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:20Z","lastTransitionTime":"2025-12-06T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.374486 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.374541 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.374568 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.374595 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.374616 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:20Z","lastTransitionTime":"2025-12-06T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.417206 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:57:20 crc kubenswrapper[4706]: E1206 13:57:20.417441 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.417410637 +0000 UTC m=+147.253203534 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.417830 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.418062 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:20 crc kubenswrapper[4706]: E1206 13:57:20.418214 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.418235 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.418407 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:20 crc kubenswrapper[4706]: E1206 13:57:20.418295 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 13:57:20 crc kubenswrapper[4706]: E1206 13:57:20.418461 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:57:20 crc kubenswrapper[4706]: E1206 13:57:20.418255 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 13:57:20 crc kubenswrapper[4706]: E1206 13:57:20.418509 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 13:57:20 crc kubenswrapper[4706]: E1206 13:57:20.418533 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:57:20 crc kubenswrapper[4706]: E1206 13:57:20.418562 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.418529248 +0000 UTC m=+147.254322145 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:57:20 crc kubenswrapper[4706]: E1206 13:57:20.418570 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 13:57:20 crc kubenswrapper[4706]: E1206 13:57:20.418594 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.418578639 +0000 UTC m=+147.254371536 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 13:57:20 crc kubenswrapper[4706]: E1206 13:57:20.418657 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.41863099 +0000 UTC m=+147.254424097 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 13:57:20 crc kubenswrapper[4706]: E1206 13:57:20.419224 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 13:57:20 crc kubenswrapper[4706]: E1206 13:57:20.419540 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.419509405 +0000 UTC m=+147.255302302 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.477947 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.477994 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.478018 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.478045 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.478069 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:20Z","lastTransitionTime":"2025-12-06T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.510681 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:20 crc kubenswrapper[4706]: E1206 13:57:20.510852 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.581425 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.581795 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.581957 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.582513 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.582691 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:20Z","lastTransitionTime":"2025-12-06T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.686619 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.687040 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.687508 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.687570 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.687591 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:20Z","lastTransitionTime":"2025-12-06T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.791641 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.791706 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.791725 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.791751 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.791769 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:20Z","lastTransitionTime":"2025-12-06T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.895060 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.895159 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.895179 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.895213 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.895236 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:20Z","lastTransitionTime":"2025-12-06T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.998411 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.998855 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.998925 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.998957 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:20 crc kubenswrapper[4706]: I1206 13:57:20.998982 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:20Z","lastTransitionTime":"2025-12-06T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.102078 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.102161 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.102186 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.102220 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.102240 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:21Z","lastTransitionTime":"2025-12-06T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.204986 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.205056 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.205074 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.205098 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.205120 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:21Z","lastTransitionTime":"2025-12-06T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.309234 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.309285 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.309349 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.309377 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.309393 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:21Z","lastTransitionTime":"2025-12-06T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.413135 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.413197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.413215 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.413239 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.413257 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:21Z","lastTransitionTime":"2025-12-06T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.510761 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.510779 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:21 crc kubenswrapper[4706]: E1206 13:57:21.510999 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.511033 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:21 crc kubenswrapper[4706]: E1206 13:57:21.511180 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:21 crc kubenswrapper[4706]: E1206 13:57:21.511403 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.515855 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.515919 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.515941 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.515967 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.515989 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:21Z","lastTransitionTime":"2025-12-06T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.618638 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.618693 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.618860 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.618889 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.620574 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:21Z","lastTransitionTime":"2025-12-06T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.723830 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.723892 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.723914 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.723959 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.723985 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:21Z","lastTransitionTime":"2025-12-06T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.827803 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.827867 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.827886 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.827912 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.827930 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:21Z","lastTransitionTime":"2025-12-06T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.930818 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.930890 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.930910 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.930937 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:21 crc kubenswrapper[4706]: I1206 13:57:21.930955 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:21Z","lastTransitionTime":"2025-12-06T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.034820 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.034887 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.034902 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.034931 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.034952 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:22Z","lastTransitionTime":"2025-12-06T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.138060 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.138140 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.138162 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.138192 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.138215 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:22Z","lastTransitionTime":"2025-12-06T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.241377 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.241455 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.241476 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.241500 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.241518 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:22Z","lastTransitionTime":"2025-12-06T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.343801 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.343864 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.343882 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.343905 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.343923 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:22Z","lastTransitionTime":"2025-12-06T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.447885 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.447949 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.447967 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.447992 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.448011 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:22Z","lastTransitionTime":"2025-12-06T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.510148 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:22 crc kubenswrapper[4706]: E1206 13:57:22.510367 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.550853 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.550902 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.550913 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.550931 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.550943 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:22Z","lastTransitionTime":"2025-12-06T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.654532 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.654572 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.654585 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.654602 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.654616 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:22Z","lastTransitionTime":"2025-12-06T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.759113 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.759228 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.759259 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.759297 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.759363 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:22Z","lastTransitionTime":"2025-12-06T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.861814 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.861895 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.861910 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.861932 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.861947 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:22Z","lastTransitionTime":"2025-12-06T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.964843 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.964908 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.964926 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.964952 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:22 crc kubenswrapper[4706]: I1206 13:57:22.964969 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:22Z","lastTransitionTime":"2025-12-06T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.067666 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.067725 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.067736 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.067753 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.067768 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:23Z","lastTransitionTime":"2025-12-06T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.170735 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.170788 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.170812 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.170841 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.170864 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:23Z","lastTransitionTime":"2025-12-06T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.274027 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.274090 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.274104 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.274123 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.274136 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:23Z","lastTransitionTime":"2025-12-06T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.377003 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.377062 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.377080 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.377102 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.377120 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:23Z","lastTransitionTime":"2025-12-06T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.480977 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.481068 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.481092 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.481126 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.481150 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:23Z","lastTransitionTime":"2025-12-06T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.511016 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.511225 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.511241 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:23 crc kubenswrapper[4706]: E1206 13:57:23.511717 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:23 crc kubenswrapper[4706]: E1206 13:57:23.511950 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:23 crc kubenswrapper[4706]: E1206 13:57:23.512218 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.533175 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.584397 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.584486 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.584505 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.584569 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.584592 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:23Z","lastTransitionTime":"2025-12-06T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.688889 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.688971 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.688996 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.689026 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.689050 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:23Z","lastTransitionTime":"2025-12-06T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.793111 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.793172 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.793189 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.793214 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.793234 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:23Z","lastTransitionTime":"2025-12-06T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.895930 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.896007 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.896030 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.896058 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.896081 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:23Z","lastTransitionTime":"2025-12-06T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.999489 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.999551 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.999573 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.999604 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:23 crc kubenswrapper[4706]: I1206 13:57:23.999630 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:23Z","lastTransitionTime":"2025-12-06T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.102718 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.102795 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.102823 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.102853 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.102876 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:24Z","lastTransitionTime":"2025-12-06T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.206420 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.206474 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.206490 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.206515 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.206533 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:24Z","lastTransitionTime":"2025-12-06T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.309265 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.309367 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.309394 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.309420 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.309445 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:24Z","lastTransitionTime":"2025-12-06T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.412963 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.413027 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.413040 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.413062 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.413076 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:24Z","lastTransitionTime":"2025-12-06T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.511078 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:24 crc kubenswrapper[4706]: E1206 13:57:24.511419 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.516269 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.516383 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.516403 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.516429 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.516447 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:24Z","lastTransitionTime":"2025-12-06T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.620030 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.620097 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.620120 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.620153 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.620171 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:24Z","lastTransitionTime":"2025-12-06T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.723273 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.723386 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.723413 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.723440 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.723458 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:24Z","lastTransitionTime":"2025-12-06T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.826127 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.826192 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.826210 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.826235 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.826255 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:24Z","lastTransitionTime":"2025-12-06T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.928859 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.928916 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.928932 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.928955 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:24 crc kubenswrapper[4706]: I1206 13:57:24.928974 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:24Z","lastTransitionTime":"2025-12-06T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.031917 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.032011 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.032052 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.032076 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.032093 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:25Z","lastTransitionTime":"2025-12-06T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.134363 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.134422 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.134443 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.134473 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.134495 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:25Z","lastTransitionTime":"2025-12-06T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.237793 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.237864 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.237883 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.237906 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.237925 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:25Z","lastTransitionTime":"2025-12-06T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.341496 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.341576 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.341597 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.341621 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.341639 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:25Z","lastTransitionTime":"2025-12-06T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.446090 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.446160 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.446176 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.446203 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.446226 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:25Z","lastTransitionTime":"2025-12-06T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.511148 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.511303 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:25 crc kubenswrapper[4706]: E1206 13:57:25.511554 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.511592 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:25 crc kubenswrapper[4706]: E1206 13:57:25.511946 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:25 crc kubenswrapper[4706]: E1206 13:57:25.512059 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.549197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.549253 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.549269 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.549291 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.549337 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:25Z","lastTransitionTime":"2025-12-06T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.652291 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.652389 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.652407 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.652432 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.652450 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:25Z","lastTransitionTime":"2025-12-06T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.755256 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.755388 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.755415 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.755445 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.755469 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:25Z","lastTransitionTime":"2025-12-06T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.858501 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.858559 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.858581 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.858609 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.858628 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:25Z","lastTransitionTime":"2025-12-06T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.961641 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.961710 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.961727 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.961751 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:25 crc kubenswrapper[4706]: I1206 13:57:25.961769 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:25Z","lastTransitionTime":"2025-12-06T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.065245 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.065358 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.065384 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.065418 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.065444 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:26Z","lastTransitionTime":"2025-12-06T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.168819 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.168880 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.168897 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.168921 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.168940 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:26Z","lastTransitionTime":"2025-12-06T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.271576 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.271639 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.271657 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.271679 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.271696 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:26Z","lastTransitionTime":"2025-12-06T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.374859 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.374917 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.374935 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.374957 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.374974 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:26Z","lastTransitionTime":"2025-12-06T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.478643 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.478690 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.478706 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.478728 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.478743 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:26Z","lastTransitionTime":"2025-12-06T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.510465 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:26 crc kubenswrapper[4706]: E1206 13:57:26.510614 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.511483 4706 scope.go:117] "RemoveContainer" containerID="aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17" Dec 06 13:57:26 crc kubenswrapper[4706]: E1206 13:57:26.511735 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-q6bgc_openshift-ovn-kubernetes(c6db0d03-fbc0-4b3b-b89d-282767c274be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.582111 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.582168 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.582190 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.582253 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.582273 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:26Z","lastTransitionTime":"2025-12-06T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.649870 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=64.649851719 podStartE2EDuration="1m4.649851719s" podCreationTimestamp="2025-12-06 13:56:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:57:26.649449928 +0000 UTC m=+89.485242805" watchObservedRunningTime="2025-12-06 13:57:26.649851719 +0000 UTC m=+89.485644586" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.683957 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=66.683940201 podStartE2EDuration="1m6.683940201s" podCreationTimestamp="2025-12-06 13:56:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:57:26.68318371 +0000 UTC m=+89.518976577" watchObservedRunningTime="2025-12-06 13:57:26.683940201 +0000 UTC m=+89.519733058" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.684890 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.684963 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.684981 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.685009 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.685026 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:26Z","lastTransitionTime":"2025-12-06T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.701399 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=70.701379198 podStartE2EDuration="1m10.701379198s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:57:26.701004068 +0000 UTC m=+89.536796955" watchObservedRunningTime="2025-12-06 13:57:26.701379198 +0000 UTC m=+89.537172095" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.744625 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podStartSLOduration=70.744596579 podStartE2EDuration="1m10.744596579s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:57:26.731474031 +0000 UTC m=+89.567266918" watchObservedRunningTime="2025-12-06 13:57:26.744596579 +0000 UTC m=+89.580389466" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.744833 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-49gn6" podStartSLOduration=70.744824976 podStartE2EDuration="1m10.744824976s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:57:26.744016184 +0000 UTC m=+89.579809071" watchObservedRunningTime="2025-12-06 13:57:26.744824976 +0000 UTC m=+89.580617863" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.758642 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=38.758614883 podStartE2EDuration="38.758614883s" podCreationTimestamp="2025-12-06 13:56:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:57:26.757920094 +0000 UTC m=+89.593713031" watchObservedRunningTime="2025-12-06 13:57:26.758614883 +0000 UTC m=+89.594407780" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.788249 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.788347 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.788378 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.788404 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.788421 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:26Z","lastTransitionTime":"2025-12-06T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.822496 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-xqlx6" podStartSLOduration=70.822469098 podStartE2EDuration="1m10.822469098s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:57:26.821672407 +0000 UTC m=+89.657465294" watchObservedRunningTime="2025-12-06 13:57:26.822469098 +0000 UTC m=+89.658261995" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.854430 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=3.854401981 podStartE2EDuration="3.854401981s" podCreationTimestamp="2025-12-06 13:57:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:57:26.854299979 +0000 UTC m=+89.690092866" watchObservedRunningTime="2025-12-06 13:57:26.854401981 +0000 UTC m=+89.690194868" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.891549 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.891603 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.891621 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.891644 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.891662 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:26Z","lastTransitionTime":"2025-12-06T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.896362 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-9rrsc" podStartSLOduration=70.896293767 podStartE2EDuration="1m10.896293767s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:57:26.894713554 +0000 UTC m=+89.730506451" watchObservedRunningTime="2025-12-06 13:57:26.896293767 +0000 UTC m=+89.732086664" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.911141 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-2cc78" podStartSLOduration=70.911115462 podStartE2EDuration="1m10.911115462s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:57:26.909444386 +0000 UTC m=+89.745237303" watchObservedRunningTime="2025-12-06 13:57:26.911115462 +0000 UTC m=+89.746908339" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.993915 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.993949 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.993960 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.993975 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:26 crc kubenswrapper[4706]: I1206 13:57:26.993985 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:26Z","lastTransitionTime":"2025-12-06T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.096193 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.096254 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.096275 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.096304 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.096367 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:27Z","lastTransitionTime":"2025-12-06T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.199694 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.199767 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.199783 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.199807 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.199825 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:27Z","lastTransitionTime":"2025-12-06T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.303207 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.303267 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.303286 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.303338 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.303357 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:27Z","lastTransitionTime":"2025-12-06T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.406475 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.406540 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.406559 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.406594 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.406619 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:27Z","lastTransitionTime":"2025-12-06T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.509477 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.509574 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.509592 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.509616 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.509635 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:27Z","lastTransitionTime":"2025-12-06T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.510383 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.510470 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:27 crc kubenswrapper[4706]: E1206 13:57:27.513620 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.514158 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:27 crc kubenswrapper[4706]: E1206 13:57:27.514448 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:27 crc kubenswrapper[4706]: E1206 13:57:27.514963 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.612666 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.612740 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.612781 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.612810 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.612831 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:27Z","lastTransitionTime":"2025-12-06T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.715724 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.715789 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.715813 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.715840 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.715864 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:27Z","lastTransitionTime":"2025-12-06T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.818585 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.818639 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.818656 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.818679 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.818695 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:27Z","lastTransitionTime":"2025-12-06T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.921952 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.922011 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.922028 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.922054 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:27 crc kubenswrapper[4706]: I1206 13:57:27.922071 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:27Z","lastTransitionTime":"2025-12-06T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.025838 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.025976 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.025993 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.026021 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.026038 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:28Z","lastTransitionTime":"2025-12-06T13:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.128359 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.128400 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.128412 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.128428 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.128439 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:28Z","lastTransitionTime":"2025-12-06T13:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.233125 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.233184 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.233201 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.233228 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.233245 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:28Z","lastTransitionTime":"2025-12-06T13:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.274142 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.274212 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.274228 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.274255 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.274273 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T13:57:28Z","lastTransitionTime":"2025-12-06T13:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.338727 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dzzz7" podStartSLOduration=71.338704352 podStartE2EDuration="1m11.338704352s" podCreationTimestamp="2025-12-06 13:56:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:57:26.925441694 +0000 UTC m=+89.761234581" watchObservedRunningTime="2025-12-06 13:57:28.338704352 +0000 UTC m=+91.174497239" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.339729 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt"] Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.340211 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.343056 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.343277 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.344455 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.345620 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.422020 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc88c17e-186e-4dc2-9b59-2d63324750d4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xdvvt\" (UID: \"dc88c17e-186e-4dc2-9b59-2d63324750d4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.422197 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc88c17e-186e-4dc2-9b59-2d63324750d4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xdvvt\" (UID: \"dc88c17e-186e-4dc2-9b59-2d63324750d4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.422365 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/dc88c17e-186e-4dc2-9b59-2d63324750d4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xdvvt\" (UID: \"dc88c17e-186e-4dc2-9b59-2d63324750d4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.422424 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/dc88c17e-186e-4dc2-9b59-2d63324750d4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xdvvt\" (UID: \"dc88c17e-186e-4dc2-9b59-2d63324750d4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.422464 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dc88c17e-186e-4dc2-9b59-2d63324750d4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xdvvt\" (UID: \"dc88c17e-186e-4dc2-9b59-2d63324750d4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.510026 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:28 crc kubenswrapper[4706]: E1206 13:57:28.510223 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.523777 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/dc88c17e-186e-4dc2-9b59-2d63324750d4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xdvvt\" (UID: \"dc88c17e-186e-4dc2-9b59-2d63324750d4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.523848 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/dc88c17e-186e-4dc2-9b59-2d63324750d4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xdvvt\" (UID: \"dc88c17e-186e-4dc2-9b59-2d63324750d4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.523888 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dc88c17e-186e-4dc2-9b59-2d63324750d4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xdvvt\" (UID: \"dc88c17e-186e-4dc2-9b59-2d63324750d4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.523950 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/dc88c17e-186e-4dc2-9b59-2d63324750d4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xdvvt\" (UID: \"dc88c17e-186e-4dc2-9b59-2d63324750d4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.523981 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc88c17e-186e-4dc2-9b59-2d63324750d4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xdvvt\" (UID: \"dc88c17e-186e-4dc2-9b59-2d63324750d4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.523982 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/dc88c17e-186e-4dc2-9b59-2d63324750d4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xdvvt\" (UID: \"dc88c17e-186e-4dc2-9b59-2d63324750d4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.524035 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc88c17e-186e-4dc2-9b59-2d63324750d4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xdvvt\" (UID: \"dc88c17e-186e-4dc2-9b59-2d63324750d4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.525927 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dc88c17e-186e-4dc2-9b59-2d63324750d4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xdvvt\" (UID: \"dc88c17e-186e-4dc2-9b59-2d63324750d4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.531002 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc88c17e-186e-4dc2-9b59-2d63324750d4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xdvvt\" (UID: \"dc88c17e-186e-4dc2-9b59-2d63324750d4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.551444 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc88c17e-186e-4dc2-9b59-2d63324750d4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xdvvt\" (UID: \"dc88c17e-186e-4dc2-9b59-2d63324750d4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" Dec 06 13:57:28 crc kubenswrapper[4706]: I1206 13:57:28.664572 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" Dec 06 13:57:28 crc kubenswrapper[4706]: W1206 13:57:28.686813 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc88c17e_186e_4dc2_9b59_2d63324750d4.slice/crio-53f7cdc76e787a77ef8d31a36ee51ff15e43e2643b1396c5564bdc746bfe0201 WatchSource:0}: Error finding container 53f7cdc76e787a77ef8d31a36ee51ff15e43e2643b1396c5564bdc746bfe0201: Status 404 returned error can't find the container with id 53f7cdc76e787a77ef8d31a36ee51ff15e43e2643b1396c5564bdc746bfe0201 Dec 06 13:57:29 crc kubenswrapper[4706]: I1206 13:57:29.094223 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" event={"ID":"dc88c17e-186e-4dc2-9b59-2d63324750d4","Type":"ContainerStarted","Data":"cc9268432703861951b04b2c32c6c2d700240d16d956c356607145cf95d3be04"} Dec 06 13:57:29 crc kubenswrapper[4706]: I1206 13:57:29.094720 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" event={"ID":"dc88c17e-186e-4dc2-9b59-2d63324750d4","Type":"ContainerStarted","Data":"53f7cdc76e787a77ef8d31a36ee51ff15e43e2643b1396c5564bdc746bfe0201"} Dec 06 13:57:29 crc kubenswrapper[4706]: I1206 13:57:29.111022 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdvvt" podStartSLOduration=73.111001735 podStartE2EDuration="1m13.111001735s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:57:29.110852191 +0000 UTC m=+91.946645078" watchObservedRunningTime="2025-12-06 13:57:29.111001735 +0000 UTC m=+91.946794582" Dec 06 13:57:29 crc kubenswrapper[4706]: I1206 13:57:29.510208 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:29 crc kubenswrapper[4706]: I1206 13:57:29.510224 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:29 crc kubenswrapper[4706]: I1206 13:57:29.510302 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:29 crc kubenswrapper[4706]: E1206 13:57:29.510429 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:29 crc kubenswrapper[4706]: E1206 13:57:29.510658 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:29 crc kubenswrapper[4706]: E1206 13:57:29.510753 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:30 crc kubenswrapper[4706]: I1206 13:57:30.510712 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:30 crc kubenswrapper[4706]: E1206 13:57:30.510929 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:31 crc kubenswrapper[4706]: I1206 13:57:31.510697 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:31 crc kubenswrapper[4706]: I1206 13:57:31.510742 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:31 crc kubenswrapper[4706]: I1206 13:57:31.510836 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:31 crc kubenswrapper[4706]: E1206 13:57:31.510997 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:31 crc kubenswrapper[4706]: E1206 13:57:31.511089 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:31 crc kubenswrapper[4706]: E1206 13:57:31.511140 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:32 crc kubenswrapper[4706]: I1206 13:57:32.510889 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:32 crc kubenswrapper[4706]: E1206 13:57:32.511359 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:33 crc kubenswrapper[4706]: I1206 13:57:33.510771 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:33 crc kubenswrapper[4706]: I1206 13:57:33.510812 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:33 crc kubenswrapper[4706]: I1206 13:57:33.511221 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:33 crc kubenswrapper[4706]: E1206 13:57:33.511466 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:33 crc kubenswrapper[4706]: E1206 13:57:33.511373 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:33 crc kubenswrapper[4706]: E1206 13:57:33.512316 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:34 crc kubenswrapper[4706]: I1206 13:57:34.510522 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:34 crc kubenswrapper[4706]: E1206 13:57:34.510746 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:35 crc kubenswrapper[4706]: I1206 13:57:35.198819 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs\") pod \"network-metrics-daemon-hhkhk\" (UID: \"5443fc82-2d54-43df-88da-043c2eb77238\") " pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:35 crc kubenswrapper[4706]: E1206 13:57:35.199021 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 13:57:35 crc kubenswrapper[4706]: E1206 13:57:35.199081 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs podName:5443fc82-2d54-43df-88da-043c2eb77238 nodeName:}" failed. No retries permitted until 2025-12-06 13:58:39.199065129 +0000 UTC m=+162.034857986 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs") pod "network-metrics-daemon-hhkhk" (UID: "5443fc82-2d54-43df-88da-043c2eb77238") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 13:57:35 crc kubenswrapper[4706]: I1206 13:57:35.510365 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:35 crc kubenswrapper[4706]: E1206 13:57:35.510531 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:35 crc kubenswrapper[4706]: I1206 13:57:35.510800 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:35 crc kubenswrapper[4706]: E1206 13:57:35.510873 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:35 crc kubenswrapper[4706]: I1206 13:57:35.511121 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:35 crc kubenswrapper[4706]: E1206 13:57:35.511188 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:36 crc kubenswrapper[4706]: I1206 13:57:36.510445 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:36 crc kubenswrapper[4706]: E1206 13:57:36.510612 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:37 crc kubenswrapper[4706]: I1206 13:57:37.510747 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:37 crc kubenswrapper[4706]: I1206 13:57:37.513073 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:37 crc kubenswrapper[4706]: E1206 13:57:37.513057 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:37 crc kubenswrapper[4706]: E1206 13:57:37.513232 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:37 crc kubenswrapper[4706]: I1206 13:57:37.513344 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:37 crc kubenswrapper[4706]: E1206 13:57:37.513535 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:38 crc kubenswrapper[4706]: I1206 13:57:38.511110 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:38 crc kubenswrapper[4706]: E1206 13:57:38.511405 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:39 crc kubenswrapper[4706]: I1206 13:57:39.510406 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:39 crc kubenswrapper[4706]: I1206 13:57:39.510366 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:39 crc kubenswrapper[4706]: E1206 13:57:39.510502 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:39 crc kubenswrapper[4706]: E1206 13:57:39.510652 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:39 crc kubenswrapper[4706]: I1206 13:57:39.510756 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:39 crc kubenswrapper[4706]: E1206 13:57:39.511069 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:39 crc kubenswrapper[4706]: I1206 13:57:39.511241 4706 scope.go:117] "RemoveContainer" containerID="aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17" Dec 06 13:57:39 crc kubenswrapper[4706]: E1206 13:57:39.511412 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-q6bgc_openshift-ovn-kubernetes(c6db0d03-fbc0-4b3b-b89d-282767c274be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" Dec 06 13:57:40 crc kubenswrapper[4706]: I1206 13:57:40.510919 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:40 crc kubenswrapper[4706]: E1206 13:57:40.511123 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:41 crc kubenswrapper[4706]: I1206 13:57:41.510677 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:41 crc kubenswrapper[4706]: I1206 13:57:41.510827 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:41 crc kubenswrapper[4706]: E1206 13:57:41.511021 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:41 crc kubenswrapper[4706]: I1206 13:57:41.511059 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:41 crc kubenswrapper[4706]: E1206 13:57:41.511240 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:41 crc kubenswrapper[4706]: E1206 13:57:41.511462 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:42 crc kubenswrapper[4706]: I1206 13:57:42.510741 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:42 crc kubenswrapper[4706]: E1206 13:57:42.510913 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:43 crc kubenswrapper[4706]: I1206 13:57:43.510356 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:43 crc kubenswrapper[4706]: E1206 13:57:43.510549 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:43 crc kubenswrapper[4706]: I1206 13:57:43.510558 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:43 crc kubenswrapper[4706]: I1206 13:57:43.510394 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:43 crc kubenswrapper[4706]: E1206 13:57:43.510665 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:43 crc kubenswrapper[4706]: E1206 13:57:43.510871 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:44 crc kubenswrapper[4706]: I1206 13:57:44.510702 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:44 crc kubenswrapper[4706]: E1206 13:57:44.511008 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:45 crc kubenswrapper[4706]: I1206 13:57:45.511145 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:45 crc kubenswrapper[4706]: I1206 13:57:45.511226 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:45 crc kubenswrapper[4706]: E1206 13:57:45.511293 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:45 crc kubenswrapper[4706]: E1206 13:57:45.511460 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:45 crc kubenswrapper[4706]: I1206 13:57:45.511536 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:45 crc kubenswrapper[4706]: E1206 13:57:45.511645 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:46 crc kubenswrapper[4706]: I1206 13:57:46.510638 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:46 crc kubenswrapper[4706]: E1206 13:57:46.510964 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:47 crc kubenswrapper[4706]: I1206 13:57:47.510713 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:47 crc kubenswrapper[4706]: I1206 13:57:47.510869 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:47 crc kubenswrapper[4706]: I1206 13:57:47.510983 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:47 crc kubenswrapper[4706]: E1206 13:57:47.512567 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:47 crc kubenswrapper[4706]: E1206 13:57:47.512784 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:47 crc kubenswrapper[4706]: E1206 13:57:47.512916 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:48 crc kubenswrapper[4706]: I1206 13:57:48.510571 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:48 crc kubenswrapper[4706]: E1206 13:57:48.510742 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:49 crc kubenswrapper[4706]: I1206 13:57:49.510745 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:49 crc kubenswrapper[4706]: E1206 13:57:49.510926 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:49 crc kubenswrapper[4706]: I1206 13:57:49.511178 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:49 crc kubenswrapper[4706]: I1206 13:57:49.510773 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:49 crc kubenswrapper[4706]: E1206 13:57:49.511726 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:49 crc kubenswrapper[4706]: E1206 13:57:49.511924 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:50 crc kubenswrapper[4706]: I1206 13:57:50.510524 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:50 crc kubenswrapper[4706]: E1206 13:57:50.510682 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:50 crc kubenswrapper[4706]: I1206 13:57:50.511601 4706 scope.go:117] "RemoveContainer" containerID="aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17" Dec 06 13:57:50 crc kubenswrapper[4706]: E1206 13:57:50.511842 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-q6bgc_openshift-ovn-kubernetes(c6db0d03-fbc0-4b3b-b89d-282767c274be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" Dec 06 13:57:51 crc kubenswrapper[4706]: I1206 13:57:51.187224 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xqlx6_d3fdd6dc-f817-486c-b0ee-18a6f2185559/kube-multus/1.log" Dec 06 13:57:51 crc kubenswrapper[4706]: I1206 13:57:51.188120 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xqlx6_d3fdd6dc-f817-486c-b0ee-18a6f2185559/kube-multus/0.log" Dec 06 13:57:51 crc kubenswrapper[4706]: I1206 13:57:51.188198 4706 generic.go:334] "Generic (PLEG): container finished" podID="d3fdd6dc-f817-486c-b0ee-18a6f2185559" containerID="fcaf233bcb76d165e238dfbafde4bcbfa99d5a083980078c23ea17e2ab19d5af" exitCode=1 Dec 06 13:57:51 crc kubenswrapper[4706]: I1206 13:57:51.188270 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xqlx6" event={"ID":"d3fdd6dc-f817-486c-b0ee-18a6f2185559","Type":"ContainerDied","Data":"fcaf233bcb76d165e238dfbafde4bcbfa99d5a083980078c23ea17e2ab19d5af"} Dec 06 13:57:51 crc kubenswrapper[4706]: I1206 13:57:51.188417 4706 scope.go:117] "RemoveContainer" containerID="db9d05c25d0d13df7f13d3ffbaec3cc1f239a1e53499f718593834b7d37eafd0" Dec 06 13:57:51 crc kubenswrapper[4706]: I1206 13:57:51.189024 4706 scope.go:117] "RemoveContainer" containerID="fcaf233bcb76d165e238dfbafde4bcbfa99d5a083980078c23ea17e2ab19d5af" Dec 06 13:57:51 crc kubenswrapper[4706]: E1206 13:57:51.189601 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-xqlx6_openshift-multus(d3fdd6dc-f817-486c-b0ee-18a6f2185559)\"" pod="openshift-multus/multus-xqlx6" podUID="d3fdd6dc-f817-486c-b0ee-18a6f2185559" Dec 06 13:57:51 crc kubenswrapper[4706]: I1206 13:57:51.510132 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:51 crc kubenswrapper[4706]: I1206 13:57:51.510249 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:51 crc kubenswrapper[4706]: E1206 13:57:51.510301 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:51 crc kubenswrapper[4706]: E1206 13:57:51.510472 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:51 crc kubenswrapper[4706]: I1206 13:57:51.510551 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:51 crc kubenswrapper[4706]: E1206 13:57:51.510666 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:52 crc kubenswrapper[4706]: I1206 13:57:52.194476 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xqlx6_d3fdd6dc-f817-486c-b0ee-18a6f2185559/kube-multus/1.log" Dec 06 13:57:52 crc kubenswrapper[4706]: I1206 13:57:52.510019 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:52 crc kubenswrapper[4706]: E1206 13:57:52.510261 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:53 crc kubenswrapper[4706]: I1206 13:57:53.510970 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:53 crc kubenswrapper[4706]: E1206 13:57:53.511216 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:53 crc kubenswrapper[4706]: I1206 13:57:53.511289 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:53 crc kubenswrapper[4706]: I1206 13:57:53.511332 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:53 crc kubenswrapper[4706]: E1206 13:57:53.511498 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:53 crc kubenswrapper[4706]: E1206 13:57:53.511627 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:54 crc kubenswrapper[4706]: I1206 13:57:54.510127 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:54 crc kubenswrapper[4706]: E1206 13:57:54.510356 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:55 crc kubenswrapper[4706]: I1206 13:57:55.510661 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:55 crc kubenswrapper[4706]: I1206 13:57:55.510728 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:55 crc kubenswrapper[4706]: I1206 13:57:55.510790 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:55 crc kubenswrapper[4706]: E1206 13:57:55.510875 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:55 crc kubenswrapper[4706]: E1206 13:57:55.511096 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:55 crc kubenswrapper[4706]: E1206 13:57:55.511248 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:56 crc kubenswrapper[4706]: I1206 13:57:56.510631 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:56 crc kubenswrapper[4706]: E1206 13:57:56.510795 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:57 crc kubenswrapper[4706]: I1206 13:57:57.510552 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:57 crc kubenswrapper[4706]: I1206 13:57:57.510567 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:57 crc kubenswrapper[4706]: I1206 13:57:57.510657 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:57 crc kubenswrapper[4706]: E1206 13:57:57.512833 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:57 crc kubenswrapper[4706]: E1206 13:57:57.512995 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:57:57 crc kubenswrapper[4706]: E1206 13:57:57.513115 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:57 crc kubenswrapper[4706]: E1206 13:57:57.536464 4706 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 06 13:57:57 crc kubenswrapper[4706]: E1206 13:57:57.618280 4706 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 13:57:58 crc kubenswrapper[4706]: I1206 13:57:58.510035 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:57:58 crc kubenswrapper[4706]: E1206 13:57:58.510223 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:57:59 crc kubenswrapper[4706]: I1206 13:57:59.510054 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:57:59 crc kubenswrapper[4706]: I1206 13:57:59.510107 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:57:59 crc kubenswrapper[4706]: I1206 13:57:59.510077 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:57:59 crc kubenswrapper[4706]: E1206 13:57:59.510253 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:57:59 crc kubenswrapper[4706]: E1206 13:57:59.510382 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:57:59 crc kubenswrapper[4706]: E1206 13:57:59.510541 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:58:00 crc kubenswrapper[4706]: I1206 13:58:00.511008 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:58:00 crc kubenswrapper[4706]: E1206 13:58:00.511190 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:58:01 crc kubenswrapper[4706]: I1206 13:58:01.510374 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:58:01 crc kubenswrapper[4706]: I1206 13:58:01.510448 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:58:01 crc kubenswrapper[4706]: I1206 13:58:01.510452 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:58:01 crc kubenswrapper[4706]: E1206 13:58:01.510605 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:58:01 crc kubenswrapper[4706]: E1206 13:58:01.510700 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:58:01 crc kubenswrapper[4706]: E1206 13:58:01.510808 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:58:02 crc kubenswrapper[4706]: I1206 13:58:02.510249 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:58:02 crc kubenswrapper[4706]: E1206 13:58:02.510829 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:58:02 crc kubenswrapper[4706]: E1206 13:58:02.620101 4706 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 13:58:03 crc kubenswrapper[4706]: I1206 13:58:03.511559 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:58:03 crc kubenswrapper[4706]: E1206 13:58:03.513173 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:58:03 crc kubenswrapper[4706]: I1206 13:58:03.512151 4706 scope.go:117] "RemoveContainer" containerID="fcaf233bcb76d165e238dfbafde4bcbfa99d5a083980078c23ea17e2ab19d5af" Dec 06 13:58:03 crc kubenswrapper[4706]: I1206 13:58:03.512192 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:58:03 crc kubenswrapper[4706]: I1206 13:58:03.511662 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:58:03 crc kubenswrapper[4706]: E1206 13:58:03.515237 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:58:03 crc kubenswrapper[4706]: E1206 13:58:03.515483 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:58:04 crc kubenswrapper[4706]: I1206 13:58:04.240879 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xqlx6_d3fdd6dc-f817-486c-b0ee-18a6f2185559/kube-multus/1.log" Dec 06 13:58:04 crc kubenswrapper[4706]: I1206 13:58:04.241146 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xqlx6" event={"ID":"d3fdd6dc-f817-486c-b0ee-18a6f2185559","Type":"ContainerStarted","Data":"eaf8244ef2489d4736fe70838f4ddad49ff130d448cd95eac58babc9d24ad22b"} Dec 06 13:58:04 crc kubenswrapper[4706]: I1206 13:58:04.510747 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:58:04 crc kubenswrapper[4706]: E1206 13:58:04.510941 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:58:05 crc kubenswrapper[4706]: I1206 13:58:05.510636 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:58:05 crc kubenswrapper[4706]: I1206 13:58:05.510716 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:58:05 crc kubenswrapper[4706]: E1206 13:58:05.510771 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:58:05 crc kubenswrapper[4706]: I1206 13:58:05.510883 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:58:05 crc kubenswrapper[4706]: E1206 13:58:05.511017 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:58:05 crc kubenswrapper[4706]: E1206 13:58:05.511083 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:58:05 crc kubenswrapper[4706]: I1206 13:58:05.512234 4706 scope.go:117] "RemoveContainer" containerID="aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17" Dec 06 13:58:06 crc kubenswrapper[4706]: I1206 13:58:06.250150 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovnkube-controller/3.log" Dec 06 13:58:06 crc kubenswrapper[4706]: I1206 13:58:06.253551 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerStarted","Data":"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7"} Dec 06 13:58:06 crc kubenswrapper[4706]: I1206 13:58:06.254141 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:58:06 crc kubenswrapper[4706]: I1206 13:58:06.287972 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podStartSLOduration=110.287951149 podStartE2EDuration="1m50.287951149s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:06.287098065 +0000 UTC m=+129.122890962" watchObservedRunningTime="2025-12-06 13:58:06.287951149 +0000 UTC m=+129.123744016" Dec 06 13:58:06 crc kubenswrapper[4706]: I1206 13:58:06.468025 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hhkhk"] Dec 06 13:58:06 crc kubenswrapper[4706]: I1206 13:58:06.468169 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:58:06 crc kubenswrapper[4706]: E1206 13:58:06.468348 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:58:06 crc kubenswrapper[4706]: I1206 13:58:06.510711 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:58:06 crc kubenswrapper[4706]: E1206 13:58:06.510953 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:58:07 crc kubenswrapper[4706]: I1206 13:58:07.510660 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:58:07 crc kubenswrapper[4706]: I1206 13:58:07.510686 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:58:07 crc kubenswrapper[4706]: E1206 13:58:07.512468 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:58:07 crc kubenswrapper[4706]: E1206 13:58:07.512625 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:58:07 crc kubenswrapper[4706]: E1206 13:58:07.621148 4706 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 13:58:08 crc kubenswrapper[4706]: I1206 13:58:08.510968 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:58:08 crc kubenswrapper[4706]: E1206 13:58:08.511112 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:58:08 crc kubenswrapper[4706]: I1206 13:58:08.510968 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:58:08 crc kubenswrapper[4706]: E1206 13:58:08.511379 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:58:09 crc kubenswrapper[4706]: I1206 13:58:09.510397 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:58:09 crc kubenswrapper[4706]: I1206 13:58:09.510464 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:58:09 crc kubenswrapper[4706]: E1206 13:58:09.510622 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:58:09 crc kubenswrapper[4706]: E1206 13:58:09.510787 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:58:10 crc kubenswrapper[4706]: I1206 13:58:10.510882 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:58:10 crc kubenswrapper[4706]: I1206 13:58:10.510882 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:58:10 crc kubenswrapper[4706]: E1206 13:58:10.511477 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:58:10 crc kubenswrapper[4706]: E1206 13:58:10.512370 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:58:11 crc kubenswrapper[4706]: I1206 13:58:11.510541 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:58:11 crc kubenswrapper[4706]: I1206 13:58:11.510566 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:58:11 crc kubenswrapper[4706]: E1206 13:58:11.510775 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 13:58:11 crc kubenswrapper[4706]: E1206 13:58:11.510863 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 13:58:12 crc kubenswrapper[4706]: I1206 13:58:12.510348 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:58:12 crc kubenswrapper[4706]: I1206 13:58:12.510366 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:58:12 crc kubenswrapper[4706]: E1206 13:58:12.510557 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhkhk" podUID="5443fc82-2d54-43df-88da-043c2eb77238" Dec 06 13:58:12 crc kubenswrapper[4706]: E1206 13:58:12.510713 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 13:58:13 crc kubenswrapper[4706]: I1206 13:58:13.510674 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:58:13 crc kubenswrapper[4706]: I1206 13:58:13.510725 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:58:13 crc kubenswrapper[4706]: I1206 13:58:13.513505 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 13:58:13 crc kubenswrapper[4706]: I1206 13:58:13.513913 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 13:58:13 crc kubenswrapper[4706]: I1206 13:58:13.514260 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 13:58:13 crc kubenswrapper[4706]: I1206 13:58:13.517047 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 13:58:14 crc kubenswrapper[4706]: I1206 13:58:14.510964 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:58:14 crc kubenswrapper[4706]: I1206 13:58:14.510964 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:58:14 crc kubenswrapper[4706]: I1206 13:58:14.514183 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 13:58:14 crc kubenswrapper[4706]: I1206 13:58:14.514348 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.211367 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.286926 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9hxp8"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.289916 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.292777 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.293518 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.293827 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.294791 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tsvsk"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.295844 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.297976 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.300279 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.302566 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.314506 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.317144 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.322388 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.322653 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.323189 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.325528 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.325707 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.326732 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.326893 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.327012 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.327153 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.327345 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.330737 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qgdc9"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.331675 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qgdc9" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.333088 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-p8xpt"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.333723 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.334099 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.334545 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.334598 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.334636 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.334927 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.335648 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.335899 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.336693 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.342288 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.342618 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.342794 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.342922 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.343093 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.343199 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.343351 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.343456 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.343587 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.343649 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.343774 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.344532 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.345508 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.346660 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.346996 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.347348 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.347481 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.347682 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.351547 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.353708 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-j4chj"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.354349 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355390 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/475504b3-7830-4226-b00c-3d35eda647fe-serving-cert\") pod \"controller-manager-879f6c89f-tsvsk\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355439 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tsvsk\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355468 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0136a364-d7ad-445a-9226-5b064e4b4527-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355493 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4c02a475-2aec-4c9c-ba4d-f20059c09682-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9hxp8\" (UID: \"4c02a475-2aec-4c9c-ba4d-f20059c09682\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355518 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a9eee3d-8221-46cf-b949-bd7e854e12ab-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qgdc9\" (UID: \"1a9eee3d-8221-46cf-b949-bd7e854e12ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qgdc9" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355567 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-client-ca\") pod \"controller-manager-879f6c89f-tsvsk\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355592 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0136a364-d7ad-445a-9226-5b064e4b4527-serving-cert\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355619 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8k5g\" (UniqueName: \"kubernetes.io/projected/475504b3-7830-4226-b00c-3d35eda647fe-kube-api-access-p8k5g\") pod \"controller-manager-879f6c89f-tsvsk\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355650 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0136a364-d7ad-445a-9226-5b064e4b4527-encryption-config\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355672 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0136a364-d7ad-445a-9226-5b064e4b4527-audit-dir\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355694 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqttw\" (UniqueName: \"kubernetes.io/projected/1a9eee3d-8221-46cf-b949-bd7e854e12ab-kube-api-access-mqttw\") pod \"openshift-apiserver-operator-796bbdcf4f-qgdc9\" (UID: \"1a9eee3d-8221-46cf-b949-bd7e854e12ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qgdc9" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355729 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a9eee3d-8221-46cf-b949-bd7e854e12ab-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qgdc9\" (UID: \"1a9eee3d-8221-46cf-b949-bd7e854e12ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qgdc9" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355754 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4c02a475-2aec-4c9c-ba4d-f20059c09682-images\") pod \"machine-api-operator-5694c8668f-9hxp8\" (UID: \"4c02a475-2aec-4c9c-ba4d-f20059c09682\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355776 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wknkv\" (UniqueName: \"kubernetes.io/projected/0136a364-d7ad-445a-9226-5b064e4b4527-kube-api-access-wknkv\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355798 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0136a364-d7ad-445a-9226-5b064e4b4527-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355823 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjw7r\" (UniqueName: \"kubernetes.io/projected/4c02a475-2aec-4c9c-ba4d-f20059c09682-kube-api-access-hjw7r\") pod \"machine-api-operator-5694c8668f-9hxp8\" (UID: \"4c02a475-2aec-4c9c-ba4d-f20059c09682\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355867 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-config\") pod \"controller-manager-879f6c89f-tsvsk\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355897 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c02a475-2aec-4c9c-ba4d-f20059c09682-config\") pod \"machine-api-operator-5694c8668f-9hxp8\" (UID: \"4c02a475-2aec-4c9c-ba4d-f20059c09682\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355925 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0136a364-d7ad-445a-9226-5b064e4b4527-audit-policies\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.355953 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0136a364-d7ad-445a-9226-5b064e4b4527-etcd-client\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.358796 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.358998 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.360486 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-595n2"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.361178 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.361701 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.361841 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-595n2" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.366568 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-9jn7l"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.367038 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9jn7l" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.368162 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.368461 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.368644 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.368940 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.369094 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.369218 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.369225 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.369566 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p85f8"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.369699 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.369722 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.370032 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.370408 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p85f8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.370680 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2nk8h"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.370925 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2nk8h" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.371512 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-r94jx"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.371757 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.371841 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.381563 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.377263 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.383445 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.384167 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.384660 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.387700 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.389242 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.389869 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.390366 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.390856 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.390953 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.391097 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.391681 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.391933 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.392107 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.392216 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.392237 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.392646 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.397489 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.398183 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.398210 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.403920 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.409902 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.435908 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp64k"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.436852 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.437562 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.437873 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-z4bwl"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.437987 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp64k" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.438107 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l55q6"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.438279 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.440169 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-t2255"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.440451 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.440784 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8h4wk"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.441012 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-ldd62"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.441338 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.441932 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.442084 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.441998 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.442026 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8h4wk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.441970 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-t2255" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.446345 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.446517 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.446982 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.447229 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.447330 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.447668 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.447771 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.447922 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.448287 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.447950 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.448523 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.447992 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.448997 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-zx8fv"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.450114 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cl6js"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.450540 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-cl6js" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.450725 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.450848 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-zx8fv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.451847 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.451854 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.451953 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.452176 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.452589 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.453009 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.453167 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.454139 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-r2t78"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.455146 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2t78" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.458966 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b9f16df2-5700-43c2-9488-78c3a9d8bcd6-metrics-certs\") pod \"router-default-5444994796-r94jx\" (UID: \"b9f16df2-5700-43c2-9488-78c3a9d8bcd6\") " pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.458995 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe23e3f5-57af-4fd2-a8bd-9364bbe39776-config\") pod \"authentication-operator-69f744f599-j4chj\" (UID: \"fe23e3f5-57af-4fd2-a8bd-9364bbe39776\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459031 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc8af660-99b4-4d61-bc9c-4580f285d6c6-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2nk8h\" (UID: \"bc8af660-99b4-4d61-bc9c-4580f285d6c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2nk8h" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459053 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx6lf\" (UniqueName: \"kubernetes.io/projected/fe23e3f5-57af-4fd2-a8bd-9364bbe39776-kube-api-access-vx6lf\") pod \"authentication-operator-69f744f599-j4chj\" (UID: \"fe23e3f5-57af-4fd2-a8bd-9364bbe39776\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459077 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0136a364-d7ad-445a-9226-5b064e4b4527-serving-cert\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459094 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-client-ca\") pod \"controller-manager-879f6c89f-tsvsk\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459111 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8k5g\" (UniqueName: \"kubernetes.io/projected/475504b3-7830-4226-b00c-3d35eda647fe-kube-api-access-p8k5g\") pod \"controller-manager-879f6c89f-tsvsk\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459164 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w54rd\" (UniqueName: \"kubernetes.io/projected/77b4e41c-830d-48c2-8d00-97b33c7e8a29-kube-api-access-w54rd\") pod \"control-plane-machine-set-operator-78cbb6b69f-p85f8\" (UID: \"77b4e41c-830d-48c2-8d00-97b33c7e8a29\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p85f8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459182 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0136a364-d7ad-445a-9226-5b064e4b4527-audit-dir\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459201 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0136a364-d7ad-445a-9226-5b064e4b4527-encryption-config\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459217 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9f16df2-5700-43c2-9488-78c3a9d8bcd6-service-ca-bundle\") pod \"router-default-5444994796-r94jx\" (UID: \"b9f16df2-5700-43c2-9488-78c3a9d8bcd6\") " pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459237 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqttw\" (UniqueName: \"kubernetes.io/projected/1a9eee3d-8221-46cf-b949-bd7e854e12ab-kube-api-access-mqttw\") pod \"openshift-apiserver-operator-796bbdcf4f-qgdc9\" (UID: \"1a9eee3d-8221-46cf-b949-bd7e854e12ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qgdc9" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459273 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe23e3f5-57af-4fd2-a8bd-9364bbe39776-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-j4chj\" (UID: \"fe23e3f5-57af-4fd2-a8bd-9364bbe39776\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459291 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a9eee3d-8221-46cf-b949-bd7e854e12ab-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qgdc9\" (UID: \"1a9eee3d-8221-46cf-b949-bd7e854e12ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qgdc9" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459321 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8xz9\" (UniqueName: \"kubernetes.io/projected/b801f158-2df3-44c7-8bfe-5caa8467e29d-kube-api-access-d8xz9\") pod \"route-controller-manager-6576b87f9c-dmfw8\" (UID: \"b801f158-2df3-44c7-8bfe-5caa8467e29d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459409 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0acf2c47-e9d7-4759-aa80-9f4ca8488899-etcd-serving-ca\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459426 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc8af660-99b4-4d61-bc9c-4580f285d6c6-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2nk8h\" (UID: \"bc8af660-99b4-4d61-bc9c-4580f285d6c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2nk8h" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459443 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4c02a475-2aec-4c9c-ba4d-f20059c09682-images\") pod \"machine-api-operator-5694c8668f-9hxp8\" (UID: \"4c02a475-2aec-4c9c-ba4d-f20059c09682\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459461 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqhbg\" (UniqueName: \"kubernetes.io/projected/b9f16df2-5700-43c2-9488-78c3a9d8bcd6-kube-api-access-kqhbg\") pod \"router-default-5444994796-r94jx\" (UID: \"b9f16df2-5700-43c2-9488-78c3a9d8bcd6\") " pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459479 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wknkv\" (UniqueName: \"kubernetes.io/projected/0136a364-d7ad-445a-9226-5b064e4b4527-kube-api-access-wknkv\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459508 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/0acf2c47-e9d7-4759-aa80-9f4ca8488899-audit\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459523 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0136a364-d7ad-445a-9226-5b064e4b4527-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459539 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0acf2c47-e9d7-4759-aa80-9f4ca8488899-config\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459553 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b9f16df2-5700-43c2-9488-78c3a9d8bcd6-stats-auth\") pod \"router-default-5444994796-r94jx\" (UID: \"b9f16df2-5700-43c2-9488-78c3a9d8bcd6\") " pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459571 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjw7r\" (UniqueName: \"kubernetes.io/projected/4c02a475-2aec-4c9c-ba4d-f20059c09682-kube-api-access-hjw7r\") pod \"machine-api-operator-5694c8668f-9hxp8\" (UID: \"4c02a475-2aec-4c9c-ba4d-f20059c09682\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459608 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b801f158-2df3-44c7-8bfe-5caa8467e29d-config\") pod \"route-controller-manager-6576b87f9c-dmfw8\" (UID: \"b801f158-2df3-44c7-8bfe-5caa8467e29d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459635 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhks2\" (UniqueName: \"kubernetes.io/projected/0acf2c47-e9d7-4759-aa80-9f4ca8488899-kube-api-access-mhks2\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459650 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe23e3f5-57af-4fd2-a8bd-9364bbe39776-service-ca-bundle\") pod \"authentication-operator-69f744f599-j4chj\" (UID: \"fe23e3f5-57af-4fd2-a8bd-9364bbe39776\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459669 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-config\") pod \"controller-manager-879f6c89f-tsvsk\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459683 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/caf8bd48-353a-428e-ba7a-1a7c40f19b23-serving-cert\") pod \"openshift-config-operator-7777fb866f-2wmlg\" (UID: \"caf8bd48-353a-428e-ba7a-1a7c40f19b23\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459707 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b38e2a11-98f9-47c3-a711-106d6aa308de-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-595n2\" (UID: \"b38e2a11-98f9-47c3-a711-106d6aa308de\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-595n2" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459791 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c02a475-2aec-4c9c-ba4d-f20059c09682-config\") pod \"machine-api-operator-5694c8668f-9hxp8\" (UID: \"4c02a475-2aec-4c9c-ba4d-f20059c09682\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459814 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0136a364-d7ad-445a-9226-5b064e4b4527-audit-policies\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459830 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0136a364-d7ad-445a-9226-5b064e4b4527-etcd-client\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459851 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b801f158-2df3-44c7-8bfe-5caa8467e29d-serving-cert\") pod \"route-controller-manager-6576b87f9c-dmfw8\" (UID: \"b801f158-2df3-44c7-8bfe-5caa8467e29d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459867 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0acf2c47-e9d7-4759-aa80-9f4ca8488899-serving-cert\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459892 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/77b4e41c-830d-48c2-8d00-97b33c7e8a29-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-p85f8\" (UID: \"77b4e41c-830d-48c2-8d00-97b33c7e8a29\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p85f8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459923 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0acf2c47-e9d7-4759-aa80-9f4ca8488899-node-pullsecrets\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459939 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7nbw\" (UniqueName: \"kubernetes.io/projected/621a48da-c965-4d74-abd0-62e119fd81c8-kube-api-access-v7nbw\") pod \"downloads-7954f5f757-9jn7l\" (UID: \"621a48da-c965-4d74-abd0-62e119fd81c8\") " pod="openshift-console/downloads-7954f5f757-9jn7l" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459954 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b9f16df2-5700-43c2-9488-78c3a9d8bcd6-default-certificate\") pod \"router-default-5444994796-r94jx\" (UID: \"b9f16df2-5700-43c2-9488-78c3a9d8bcd6\") " pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459970 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe23e3f5-57af-4fd2-a8bd-9364bbe39776-serving-cert\") pod \"authentication-operator-69f744f599-j4chj\" (UID: \"fe23e3f5-57af-4fd2-a8bd-9364bbe39776\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.459985 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0acf2c47-e9d7-4759-aa80-9f4ca8488899-etcd-client\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.460013 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/0acf2c47-e9d7-4759-aa80-9f4ca8488899-image-import-ca\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.460042 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bhm9\" (UniqueName: \"kubernetes.io/projected/caf8bd48-353a-428e-ba7a-1a7c40f19b23-kube-api-access-6bhm9\") pod \"openshift-config-operator-7777fb866f-2wmlg\" (UID: \"caf8bd48-353a-428e-ba7a-1a7c40f19b23\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.460058 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b801f158-2df3-44c7-8bfe-5caa8467e29d-client-ca\") pod \"route-controller-manager-6576b87f9c-dmfw8\" (UID: \"b801f158-2df3-44c7-8bfe-5caa8467e29d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.460081 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0acf2c47-e9d7-4759-aa80-9f4ca8488899-trusted-ca-bundle\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.460095 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9kn2\" (UniqueName: \"kubernetes.io/projected/bc8af660-99b4-4d61-bc9c-4580f285d6c6-kube-api-access-f9kn2\") pod \"openshift-controller-manager-operator-756b6f6bc6-2nk8h\" (UID: \"bc8af660-99b4-4d61-bc9c-4580f285d6c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2nk8h" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.460176 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/475504b3-7830-4226-b00c-3d35eda647fe-serving-cert\") pod \"controller-manager-879f6c89f-tsvsk\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.460191 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tsvsk\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.460216 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0136a364-d7ad-445a-9226-5b064e4b4527-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.460232 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4c02a475-2aec-4c9c-ba4d-f20059c09682-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9hxp8\" (UID: \"4c02a475-2aec-4c9c-ba4d-f20059c09682\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.460257 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/caf8bd48-353a-428e-ba7a-1a7c40f19b23-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2wmlg\" (UID: \"caf8bd48-353a-428e-ba7a-1a7c40f19b23\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.460273 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0acf2c47-e9d7-4759-aa80-9f4ca8488899-audit-dir\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.460287 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5hnr\" (UniqueName: \"kubernetes.io/projected/b38e2a11-98f9-47c3-a711-106d6aa308de-kube-api-access-k5hnr\") pod \"cluster-samples-operator-665b6dd947-595n2\" (UID: \"b38e2a11-98f9-47c3-a711-106d6aa308de\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-595n2" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.462552 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a9eee3d-8221-46cf-b949-bd7e854e12ab-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qgdc9\" (UID: \"1a9eee3d-8221-46cf-b949-bd7e854e12ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qgdc9" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.463027 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.463735 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.464698 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dp75h"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.466341 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.466361 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0136a364-d7ad-445a-9226-5b064e4b4527-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.466581 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.466689 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-client-ca\") pod \"controller-manager-879f6c89f-tsvsk\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.466778 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.466873 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0136a364-d7ad-445a-9226-5b064e4b4527-audit-dir\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.467139 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c02a475-2aec-4c9c-ba4d-f20059c09682-config\") pod \"machine-api-operator-5694c8668f-9hxp8\" (UID: \"4c02a475-2aec-4c9c-ba4d-f20059c09682\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.467200 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.467422 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0136a364-d7ad-445a-9226-5b064e4b4527-audit-policies\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.467424 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0acf2c47-e9d7-4759-aa80-9f4ca8488899-encryption-config\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.467799 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-config\") pod \"controller-manager-879f6c89f-tsvsk\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.468332 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.469492 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0136a364-d7ad-445a-9226-5b064e4b4527-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.469871 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.470372 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-97565"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.470745 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tsvsk\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.471004 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4c02a475-2aec-4c9c-ba4d-f20059c09682-images\") pod \"machine-api-operator-5694c8668f-9hxp8\" (UID: \"4c02a475-2aec-4c9c-ba4d-f20059c09682\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.471121 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9hxp8"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.471178 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-97565" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.471187 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a9eee3d-8221-46cf-b949-bd7e854e12ab-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qgdc9\" (UID: \"1a9eee3d-8221-46cf-b949-bd7e854e12ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qgdc9" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.471240 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.471260 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dp75h" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.488848 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a9eee3d-8221-46cf-b949-bd7e854e12ab-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qgdc9\" (UID: \"1a9eee3d-8221-46cf-b949-bd7e854e12ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qgdc9" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.489545 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.490002 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.490333 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4c02a475-2aec-4c9c-ba4d-f20059c09682-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9hxp8\" (UID: \"4c02a475-2aec-4c9c-ba4d-f20059c09682\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.490714 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0136a364-d7ad-445a-9226-5b064e4b4527-serving-cert\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.491073 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0136a364-d7ad-445a-9226-5b064e4b4527-etcd-client\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.491335 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.491735 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/475504b3-7830-4226-b00c-3d35eda647fe-serving-cert\") pod \"controller-manager-879f6c89f-tsvsk\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.491990 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0136a364-d7ad-445a-9226-5b064e4b4527-encryption-config\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.498214 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.500714 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.501652 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.503568 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.505754 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.506895 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-k9rh4"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.510556 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-k9rh4" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.512100 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.512581 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.517565 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cvtpw"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.518298 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.518400 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.519148 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rlczt"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.519230 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.519822 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tsvsk"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.519912 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-j2hql"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.519879 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.521025 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.521123 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-j2hql" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.521637 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qgdc9"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.521663 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-nngbv"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.521838 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.522264 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-p8xpt"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.522432 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-nngbv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.522993 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7xn8m"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.523984 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.524080 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.526586 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p85f8"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.527636 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-j4chj"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.528646 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-595n2"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.529629 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.530704 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.531688 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l55q6"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.532770 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp64k"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.533823 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9jn7l"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.534806 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-r2t78"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.535826 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.536876 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dp75h"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.537880 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.538015 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.538997 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ldd62"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.540444 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.541638 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2nk8h"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.542883 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-kdfk5"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.543345 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kdfk5" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.543971 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-m9fbd"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.544729 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.544804 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-m9fbd" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.547285 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.548804 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-k9rh4"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.551523 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cl6js"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.552203 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.564663 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-97565"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.567990 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rlczt"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569003 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b801f158-2df3-44c7-8bfe-5caa8467e29d-serving-cert\") pod \"route-controller-manager-6576b87f9c-dmfw8\" (UID: \"b801f158-2df3-44c7-8bfe-5caa8467e29d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569053 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0acf2c47-e9d7-4759-aa80-9f4ca8488899-node-pullsecrets\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569085 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0acf2c47-e9d7-4759-aa80-9f4ca8488899-serving-cert\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569108 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/77b4e41c-830d-48c2-8d00-97b33c7e8a29-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-p85f8\" (UID: \"77b4e41c-830d-48c2-8d00-97b33c7e8a29\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p85f8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569151 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzjxf\" (UniqueName: \"kubernetes.io/projected/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-kube-api-access-tzjxf\") pod \"marketplace-operator-79b997595-cvtpw\" (UID: \"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569185 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b9f16df2-5700-43c2-9488-78c3a9d8bcd6-default-certificate\") pod \"router-default-5444994796-r94jx\" (UID: \"b9f16df2-5700-43c2-9488-78c3a9d8bcd6\") " pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569207 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0acf2c47-e9d7-4759-aa80-9f4ca8488899-node-pullsecrets\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569218 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8de762a2-b5b9-4e05-aedc-9ba4d05ee95f-config\") pod \"kube-controller-manager-operator-78b949d7b-8h4wk\" (UID: \"8de762a2-b5b9-4e05-aedc-9ba4d05ee95f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8h4wk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569419 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f06c4ffb-6d3f-4189-8e4e-4d5233c41faa-config\") pod \"service-ca-operator-777779d784-j2hql\" (UID: \"f06c4ffb-6d3f-4189-8e4e-4d5233c41faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-j2hql" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569467 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cvtpw\" (UID: \"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569546 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7nbw\" (UniqueName: \"kubernetes.io/projected/621a48da-c965-4d74-abd0-62e119fd81c8-kube-api-access-v7nbw\") pod \"downloads-7954f5f757-9jn7l\" (UID: \"621a48da-c965-4d74-abd0-62e119fd81c8\") " pod="openshift-console/downloads-7954f5f757-9jn7l" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569639 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe23e3f5-57af-4fd2-a8bd-9364bbe39776-serving-cert\") pod \"authentication-operator-69f744f599-j4chj\" (UID: \"fe23e3f5-57af-4fd2-a8bd-9364bbe39776\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569704 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cc40c37-833b-439a-bfd3-b6aa63b8c033-config\") pod \"machine-approver-56656f9798-tvwsv\" (UID: \"5cc40c37-833b-439a-bfd3-b6aa63b8c033\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569749 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v26w\" (UniqueName: \"kubernetes.io/projected/9f1fce81-bf5b-48df-a92a-e95e82b5ef54-kube-api-access-8v26w\") pod \"kube-storage-version-migrator-operator-b67b599dd-6hsl2\" (UID: \"9f1fce81-bf5b-48df-a92a-e95e82b5ef54\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569802 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bhm9\" (UniqueName: \"kubernetes.io/projected/caf8bd48-353a-428e-ba7a-1a7c40f19b23-kube-api-access-6bhm9\") pod \"openshift-config-operator-7777fb866f-2wmlg\" (UID: \"caf8bd48-353a-428e-ba7a-1a7c40f19b23\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569839 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0acf2c47-e9d7-4759-aa80-9f4ca8488899-etcd-client\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569930 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/0acf2c47-e9d7-4759-aa80-9f4ca8488899-image-import-ca\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.569996 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f65fc54c-f0ca-40dc-a8bc-9084483d2d11-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gfkh4\" (UID: \"f65fc54c-f0ca-40dc-a8bc-9084483d2d11\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570052 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2fab8c56-bd27-485d-8216-c219effe9dcc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp64k\" (UID: \"2fab8c56-bd27-485d-8216-c219effe9dcc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp64k" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570135 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz264\" (UniqueName: \"kubernetes.io/projected/f65fc54c-f0ca-40dc-a8bc-9084483d2d11-kube-api-access-lz264\") pod \"olm-operator-6b444d44fb-gfkh4\" (UID: \"f65fc54c-f0ca-40dc-a8bc-9084483d2d11\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570180 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/27733066-fbcd-474e-9af2-d2bd168533c4-registration-dir\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570220 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b801f158-2df3-44c7-8bfe-5caa8467e29d-client-ca\") pod \"route-controller-manager-6576b87f9c-dmfw8\" (UID: \"b801f158-2df3-44c7-8bfe-5caa8467e29d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570281 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0acf2c47-e9d7-4759-aa80-9f4ca8488899-trusted-ca-bundle\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570355 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9kn2\" (UniqueName: \"kubernetes.io/projected/bc8af660-99b4-4d61-bc9c-4580f285d6c6-kube-api-access-f9kn2\") pod \"openshift-controller-manager-operator-756b6f6bc6-2nk8h\" (UID: \"bc8af660-99b4-4d61-bc9c-4580f285d6c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2nk8h" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570410 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk47x\" (UniqueName: \"kubernetes.io/projected/f06c4ffb-6d3f-4189-8e4e-4d5233c41faa-kube-api-access-tk47x\") pod \"service-ca-operator-777779d784-j2hql\" (UID: \"f06c4ffb-6d3f-4189-8e4e-4d5233c41faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-j2hql" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570448 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4z9d\" (UniqueName: \"kubernetes.io/projected/6b5dec00-19df-4fb7-883e-6941ebe98d0a-kube-api-access-l4z9d\") pod \"machine-config-server-nngbv\" (UID: \"6b5dec00-19df-4fb7-883e-6941ebe98d0a\") " pod="openshift-machine-config-operator/machine-config-server-nngbv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570498 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/caf8bd48-353a-428e-ba7a-1a7c40f19b23-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2wmlg\" (UID: \"caf8bd48-353a-428e-ba7a-1a7c40f19b23\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570538 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0acf2c47-e9d7-4759-aa80-9f4ca8488899-encryption-config\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570571 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0acf2c47-e9d7-4759-aa80-9f4ca8488899-audit-dir\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570620 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5hnr\" (UniqueName: \"kubernetes.io/projected/b38e2a11-98f9-47c3-a711-106d6aa308de-kube-api-access-k5hnr\") pod \"cluster-samples-operator-665b6dd947-595n2\" (UID: \"b38e2a11-98f9-47c3-a711-106d6aa308de\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-595n2" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570656 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f65fc54c-f0ca-40dc-a8bc-9084483d2d11-srv-cert\") pod \"olm-operator-6b444d44fb-gfkh4\" (UID: \"f65fc54c-f0ca-40dc-a8bc-9084483d2d11\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570687 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr8lx\" (UniqueName: \"kubernetes.io/projected/f6f43404-a6e1-49ee-bbef-9e027779cef5-kube-api-access-vr8lx\") pod \"ingress-operator-5b745b69d9-rpnhr\" (UID: \"f6f43404-a6e1-49ee-bbef-9e027779cef5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570745 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r24l7\" (UniqueName: \"kubernetes.io/projected/d2b3bc37-36b2-473e-ae10-b0bdc30f2ef2-kube-api-access-r24l7\") pod \"migrator-59844c95c7-r2t78\" (UID: \"d2b3bc37-36b2-473e-ae10-b0bdc30f2ef2\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2t78" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570778 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc8af660-99b4-4d61-bc9c-4580f285d6c6-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2nk8h\" (UID: \"bc8af660-99b4-4d61-bc9c-4580f285d6c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2nk8h" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570811 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b9f16df2-5700-43c2-9488-78c3a9d8bcd6-metrics-certs\") pod \"router-default-5444994796-r94jx\" (UID: \"b9f16df2-5700-43c2-9488-78c3a9d8bcd6\") " pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570842 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe23e3f5-57af-4fd2-a8bd-9364bbe39776-config\") pod \"authentication-operator-69f744f599-j4chj\" (UID: \"fe23e3f5-57af-4fd2-a8bd-9364bbe39776\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570875 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppmbk\" (UniqueName: \"kubernetes.io/projected/5cc40c37-833b-439a-bfd3-b6aa63b8c033-kube-api-access-ppmbk\") pod \"machine-approver-56656f9798-tvwsv\" (UID: \"5cc40c37-833b-439a-bfd3-b6aa63b8c033\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.570920 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f6f43404-a6e1-49ee-bbef-9e027779cef5-trusted-ca\") pod \"ingress-operator-5b745b69d9-rpnhr\" (UID: \"f6f43404-a6e1-49ee-bbef-9e027779cef5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.571028 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/27733066-fbcd-474e-9af2-d2bd168533c4-socket-dir\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.571069 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlpnb\" (UniqueName: \"kubernetes.io/projected/27733066-fbcd-474e-9af2-d2bd168533c4-kube-api-access-dlpnb\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.571114 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx6lf\" (UniqueName: \"kubernetes.io/projected/fe23e3f5-57af-4fd2-a8bd-9364bbe39776-kube-api-access-vx6lf\") pod \"authentication-operator-69f744f599-j4chj\" (UID: \"fe23e3f5-57af-4fd2-a8bd-9364bbe39776\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.571150 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb4q6\" (UniqueName: \"kubernetes.io/projected/56c2d16b-fba2-4c54-95fb-71e7676b4445-kube-api-access-wb4q6\") pod \"service-ca-9c57cc56f-k9rh4\" (UID: \"56c2d16b-fba2-4c54-95fb-71e7676b4445\") " pod="openshift-service-ca/service-ca-9c57cc56f-k9rh4" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.571193 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/27733066-fbcd-474e-9af2-d2bd168533c4-plugins-dir\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.571275 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5cc40c37-833b-439a-bfd3-b6aa63b8c033-machine-approver-tls\") pod \"machine-approver-56656f9798-tvwsv\" (UID: \"5cc40c37-833b-439a-bfd3-b6aa63b8c033\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.571570 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b801f158-2df3-44c7-8bfe-5caa8467e29d-client-ca\") pod \"route-controller-manager-6576b87f9c-dmfw8\" (UID: \"b801f158-2df3-44c7-8bfe-5caa8467e29d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.572764 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe23e3f5-57af-4fd2-a8bd-9364bbe39776-serving-cert\") pod \"authentication-operator-69f744f599-j4chj\" (UID: \"fe23e3f5-57af-4fd2-a8bd-9364bbe39776\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.573103 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/77b4e41c-830d-48c2-8d00-97b33c7e8a29-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-p85f8\" (UID: \"77b4e41c-830d-48c2-8d00-97b33c7e8a29\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p85f8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.573360 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f1fce81-bf5b-48df-a92a-e95e82b5ef54-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6hsl2\" (UID: \"9f1fce81-bf5b-48df-a92a-e95e82b5ef54\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.573420 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fab8c56-bd27-485d-8216-c219effe9dcc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp64k\" (UID: \"2fab8c56-bd27-485d-8216-c219effe9dcc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp64k" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.573622 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0acf2c47-e9d7-4759-aa80-9f4ca8488899-audit-dir\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.573635 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0acf2c47-e9d7-4759-aa80-9f4ca8488899-etcd-client\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.573645 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w54rd\" (UniqueName: \"kubernetes.io/projected/77b4e41c-830d-48c2-8d00-97b33c7e8a29-kube-api-access-w54rd\") pod \"control-plane-machine-set-operator-78cbb6b69f-p85f8\" (UID: \"77b4e41c-830d-48c2-8d00-97b33c7e8a29\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p85f8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.573762 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/27733066-fbcd-474e-9af2-d2bd168533c4-mountpoint-dir\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.574037 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/caf8bd48-353a-428e-ba7a-1a7c40f19b23-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2wmlg\" (UID: \"caf8bd48-353a-428e-ba7a-1a7c40f19b23\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.574047 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/0acf2c47-e9d7-4759-aa80-9f4ca8488899-image-import-ca\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.574488 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe23e3f5-57af-4fd2-a8bd-9364bbe39776-config\") pod \"authentication-operator-69f744f599-j4chj\" (UID: \"fe23e3f5-57af-4fd2-a8bd-9364bbe39776\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.574581 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.574638 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0acf2c47-e9d7-4759-aa80-9f4ca8488899-trusted-ca-bundle\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.574910 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/56c2d16b-fba2-4c54-95fb-71e7676b4445-signing-cabundle\") pod \"service-ca-9c57cc56f-k9rh4\" (UID: \"56c2d16b-fba2-4c54-95fb-71e7676b4445\") " pod="openshift-service-ca/service-ca-9c57cc56f-k9rh4" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.575177 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9f16df2-5700-43c2-9488-78c3a9d8bcd6-service-ca-bundle\") pod \"router-default-5444994796-r94jx\" (UID: \"b9f16df2-5700-43c2-9488-78c3a9d8bcd6\") " pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.575242 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6b5dec00-19df-4fb7-883e-6941ebe98d0a-certs\") pod \"machine-config-server-nngbv\" (UID: \"6b5dec00-19df-4fb7-883e-6941ebe98d0a\") " pod="openshift-machine-config-operator/machine-config-server-nngbv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.575503 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f6f43404-a6e1-49ee-bbef-9e027779cef5-bound-sa-token\") pod \"ingress-operator-5b745b69d9-rpnhr\" (UID: \"f6f43404-a6e1-49ee-bbef-9e027779cef5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.575569 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fab8c56-bd27-485d-8216-c219effe9dcc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp64k\" (UID: \"2fab8c56-bd27-485d-8216-c219effe9dcc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp64k" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.575650 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc8af660-99b4-4d61-bc9c-4580f285d6c6-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2nk8h\" (UID: \"bc8af660-99b4-4d61-bc9c-4580f285d6c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2nk8h" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.575733 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe23e3f5-57af-4fd2-a8bd-9364bbe39776-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-j4chj\" (UID: \"fe23e3f5-57af-4fd2-a8bd-9364bbe39776\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.575809 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f6f43404-a6e1-49ee-bbef-9e027779cef5-metrics-tls\") pod \"ingress-operator-5b745b69d9-rpnhr\" (UID: \"f6f43404-a6e1-49ee-bbef-9e027779cef5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.575896 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8xz9\" (UniqueName: \"kubernetes.io/projected/b801f158-2df3-44c7-8bfe-5caa8467e29d-kube-api-access-d8xz9\") pod \"route-controller-manager-6576b87f9c-dmfw8\" (UID: \"b801f158-2df3-44c7-8bfe-5caa8467e29d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.575973 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0acf2c47-e9d7-4759-aa80-9f4ca8488899-encryption-config\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.575977 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0acf2c47-e9d7-4759-aa80-9f4ca8488899-etcd-serving-ca\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576064 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/0acf2c47-e9d7-4759-aa80-9f4ca8488899-audit\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576117 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqhbg\" (UniqueName: \"kubernetes.io/projected/b9f16df2-5700-43c2-9488-78c3a9d8bcd6-kube-api-access-kqhbg\") pod \"router-default-5444994796-r94jx\" (UID: \"b9f16df2-5700-43c2-9488-78c3a9d8bcd6\") " pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576155 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8de762a2-b5b9-4e05-aedc-9ba4d05ee95f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8h4wk\" (UID: \"8de762a2-b5b9-4e05-aedc-9ba4d05ee95f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8h4wk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576210 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0acf2c47-e9d7-4759-aa80-9f4ca8488899-config\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576240 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b9f16df2-5700-43c2-9488-78c3a9d8bcd6-stats-auth\") pod \"router-default-5444994796-r94jx\" (UID: \"b9f16df2-5700-43c2-9488-78c3a9d8bcd6\") " pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576279 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cvtpw\" (UID: \"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576281 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b9f16df2-5700-43c2-9488-78c3a9d8bcd6-default-certificate\") pod \"router-default-5444994796-r94jx\" (UID: \"b9f16df2-5700-43c2-9488-78c3a9d8bcd6\") " pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576323 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/27733066-fbcd-474e-9af2-d2bd168533c4-csi-data-dir\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576349 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8de762a2-b5b9-4e05-aedc-9ba4d05ee95f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8h4wk\" (UID: \"8de762a2-b5b9-4e05-aedc-9ba4d05ee95f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8h4wk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576400 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/56c2d16b-fba2-4c54-95fb-71e7676b4445-signing-key\") pod \"service-ca-9c57cc56f-k9rh4\" (UID: \"56c2d16b-fba2-4c54-95fb-71e7676b4445\") " pod="openshift-service-ca/service-ca-9c57cc56f-k9rh4" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576427 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b801f158-2df3-44c7-8bfe-5caa8467e29d-config\") pod \"route-controller-manager-6576b87f9c-dmfw8\" (UID: \"b801f158-2df3-44c7-8bfe-5caa8467e29d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576431 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc8af660-99b4-4d61-bc9c-4580f285d6c6-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2nk8h\" (UID: \"bc8af660-99b4-4d61-bc9c-4580f285d6c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2nk8h" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576450 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhks2\" (UniqueName: \"kubernetes.io/projected/0acf2c47-e9d7-4759-aa80-9f4ca8488899-kube-api-access-mhks2\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576464 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0acf2c47-e9d7-4759-aa80-9f4ca8488899-serving-cert\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576492 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe23e3f5-57af-4fd2-a8bd-9364bbe39776-service-ca-bundle\") pod \"authentication-operator-69f744f599-j4chj\" (UID: \"fe23e3f5-57af-4fd2-a8bd-9364bbe39776\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576517 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f1fce81-bf5b-48df-a92a-e95e82b5ef54-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6hsl2\" (UID: \"9f1fce81-bf5b-48df-a92a-e95e82b5ef54\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576559 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/caf8bd48-353a-428e-ba7a-1a7c40f19b23-serving-cert\") pod \"openshift-config-operator-7777fb866f-2wmlg\" (UID: \"caf8bd48-353a-428e-ba7a-1a7c40f19b23\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576611 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6b5dec00-19df-4fb7-883e-6941ebe98d0a-node-bootstrap-token\") pod \"machine-config-server-nngbv\" (UID: \"6b5dec00-19df-4fb7-883e-6941ebe98d0a\") " pod="openshift-machine-config-operator/machine-config-server-nngbv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576648 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5cc40c37-833b-439a-bfd3-b6aa63b8c033-auth-proxy-config\") pod \"machine-approver-56656f9798-tvwsv\" (UID: \"5cc40c37-833b-439a-bfd3-b6aa63b8c033\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576667 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f06c4ffb-6d3f-4189-8e4e-4d5233c41faa-serving-cert\") pod \"service-ca-operator-777779d784-j2hql\" (UID: \"f06c4ffb-6d3f-4189-8e4e-4d5233c41faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-j2hql" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576692 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b38e2a11-98f9-47c3-a711-106d6aa308de-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-595n2\" (UID: \"b38e2a11-98f9-47c3-a711-106d6aa308de\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-595n2" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576785 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0acf2c47-e9d7-4759-aa80-9f4ca8488899-etcd-serving-ca\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576957 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe23e3f5-57af-4fd2-a8bd-9364bbe39776-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-j4chj\" (UID: \"fe23e3f5-57af-4fd2-a8bd-9364bbe39776\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.576692 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9f16df2-5700-43c2-9488-78c3a9d8bcd6-service-ca-bundle\") pod \"router-default-5444994796-r94jx\" (UID: \"b9f16df2-5700-43c2-9488-78c3a9d8bcd6\") " pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.577155 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/0acf2c47-e9d7-4759-aa80-9f4ca8488899-audit\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.577444 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc8af660-99b4-4d61-bc9c-4580f285d6c6-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2nk8h\" (UID: \"bc8af660-99b4-4d61-bc9c-4580f285d6c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2nk8h" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.577697 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe23e3f5-57af-4fd2-a8bd-9364bbe39776-service-ca-bundle\") pod \"authentication-operator-69f744f599-j4chj\" (UID: \"fe23e3f5-57af-4fd2-a8bd-9364bbe39776\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.577767 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0acf2c47-e9d7-4759-aa80-9f4ca8488899-config\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.577944 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b801f158-2df3-44c7-8bfe-5caa8467e29d-config\") pod \"route-controller-manager-6576b87f9c-dmfw8\" (UID: \"b801f158-2df3-44c7-8bfe-5caa8467e29d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.579385 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.579421 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b9f16df2-5700-43c2-9488-78c3a9d8bcd6-metrics-certs\") pod \"router-default-5444994796-r94jx\" (UID: \"b9f16df2-5700-43c2-9488-78c3a9d8bcd6\") " pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.579997 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b801f158-2df3-44c7-8bfe-5caa8467e29d-serving-cert\") pod \"route-controller-manager-6576b87f9c-dmfw8\" (UID: \"b801f158-2df3-44c7-8bfe-5caa8467e29d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.580128 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b9f16df2-5700-43c2-9488-78c3a9d8bcd6-stats-auth\") pod \"router-default-5444994796-r94jx\" (UID: \"b9f16df2-5700-43c2-9488-78c3a9d8bcd6\") " pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.580197 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/caf8bd48-353a-428e-ba7a-1a7c40f19b23-serving-cert\") pod \"openshift-config-operator-7777fb866f-2wmlg\" (UID: \"caf8bd48-353a-428e-ba7a-1a7c40f19b23\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.581046 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b38e2a11-98f9-47c3-a711-106d6aa308de-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-595n2\" (UID: \"b38e2a11-98f9-47c3-a711-106d6aa308de\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-595n2" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.581444 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8h4wk"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.583382 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-z4bwl"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.584431 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-t2255"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.586196 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-zx8fv"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.587597 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.588891 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cvtpw"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.591288 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.594928 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.594982 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kdfk5"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.596684 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.598269 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7xn8m"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.598321 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-m9fbd"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.599735 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-j2hql"] Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.611851 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.644748 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.652731 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.672902 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.677689 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/27733066-fbcd-474e-9af2-d2bd168533c4-mountpoint-dir\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.677720 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/56c2d16b-fba2-4c54-95fb-71e7676b4445-signing-cabundle\") pod \"service-ca-9c57cc56f-k9rh4\" (UID: \"56c2d16b-fba2-4c54-95fb-71e7676b4445\") " pod="openshift-service-ca/service-ca-9c57cc56f-k9rh4" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.677748 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6b5dec00-19df-4fb7-883e-6941ebe98d0a-certs\") pod \"machine-config-server-nngbv\" (UID: \"6b5dec00-19df-4fb7-883e-6941ebe98d0a\") " pod="openshift-machine-config-operator/machine-config-server-nngbv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.677778 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f6f43404-a6e1-49ee-bbef-9e027779cef5-bound-sa-token\") pod \"ingress-operator-5b745b69d9-rpnhr\" (UID: \"f6f43404-a6e1-49ee-bbef-9e027779cef5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.677795 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fab8c56-bd27-485d-8216-c219effe9dcc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp64k\" (UID: \"2fab8c56-bd27-485d-8216-c219effe9dcc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp64k" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.677812 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f6f43404-a6e1-49ee-bbef-9e027779cef5-metrics-tls\") pod \"ingress-operator-5b745b69d9-rpnhr\" (UID: \"f6f43404-a6e1-49ee-bbef-9e027779cef5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.677830 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/27733066-fbcd-474e-9af2-d2bd168533c4-mountpoint-dir\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.677846 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8de762a2-b5b9-4e05-aedc-9ba4d05ee95f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8h4wk\" (UID: \"8de762a2-b5b9-4e05-aedc-9ba4d05ee95f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8h4wk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.677868 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cvtpw\" (UID: \"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.677888 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/27733066-fbcd-474e-9af2-d2bd168533c4-csi-data-dir\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.677903 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8de762a2-b5b9-4e05-aedc-9ba4d05ee95f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8h4wk\" (UID: \"8de762a2-b5b9-4e05-aedc-9ba4d05ee95f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8h4wk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.677927 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/56c2d16b-fba2-4c54-95fb-71e7676b4445-signing-key\") pod \"service-ca-9c57cc56f-k9rh4\" (UID: \"56c2d16b-fba2-4c54-95fb-71e7676b4445\") " pod="openshift-service-ca/service-ca-9c57cc56f-k9rh4" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.677947 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f1fce81-bf5b-48df-a92a-e95e82b5ef54-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6hsl2\" (UID: \"9f1fce81-bf5b-48df-a92a-e95e82b5ef54\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.677972 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6b5dec00-19df-4fb7-883e-6941ebe98d0a-node-bootstrap-token\") pod \"machine-config-server-nngbv\" (UID: \"6b5dec00-19df-4fb7-883e-6941ebe98d0a\") " pod="openshift-machine-config-operator/machine-config-server-nngbv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.677981 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/27733066-fbcd-474e-9af2-d2bd168533c4-csi-data-dir\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.677987 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5cc40c37-833b-439a-bfd3-b6aa63b8c033-auth-proxy-config\") pod \"machine-approver-56656f9798-tvwsv\" (UID: \"5cc40c37-833b-439a-bfd3-b6aa63b8c033\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678024 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f06c4ffb-6d3f-4189-8e4e-4d5233c41faa-serving-cert\") pod \"service-ca-operator-777779d784-j2hql\" (UID: \"f06c4ffb-6d3f-4189-8e4e-4d5233c41faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-j2hql" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678055 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzjxf\" (UniqueName: \"kubernetes.io/projected/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-kube-api-access-tzjxf\") pod \"marketplace-operator-79b997595-cvtpw\" (UID: \"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678081 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8de762a2-b5b9-4e05-aedc-9ba4d05ee95f-config\") pod \"kube-controller-manager-operator-78b949d7b-8h4wk\" (UID: \"8de762a2-b5b9-4e05-aedc-9ba4d05ee95f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8h4wk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678097 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f06c4ffb-6d3f-4189-8e4e-4d5233c41faa-config\") pod \"service-ca-operator-777779d784-j2hql\" (UID: \"f06c4ffb-6d3f-4189-8e4e-4d5233c41faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-j2hql" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678114 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cvtpw\" (UID: \"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678129 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cc40c37-833b-439a-bfd3-b6aa63b8c033-config\") pod \"machine-approver-56656f9798-tvwsv\" (UID: \"5cc40c37-833b-439a-bfd3-b6aa63b8c033\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678146 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v26w\" (UniqueName: \"kubernetes.io/projected/9f1fce81-bf5b-48df-a92a-e95e82b5ef54-kube-api-access-8v26w\") pod \"kube-storage-version-migrator-operator-b67b599dd-6hsl2\" (UID: \"9f1fce81-bf5b-48df-a92a-e95e82b5ef54\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678174 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f65fc54c-f0ca-40dc-a8bc-9084483d2d11-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gfkh4\" (UID: \"f65fc54c-f0ca-40dc-a8bc-9084483d2d11\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678195 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2fab8c56-bd27-485d-8216-c219effe9dcc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp64k\" (UID: \"2fab8c56-bd27-485d-8216-c219effe9dcc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp64k" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678215 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz264\" (UniqueName: \"kubernetes.io/projected/f65fc54c-f0ca-40dc-a8bc-9084483d2d11-kube-api-access-lz264\") pod \"olm-operator-6b444d44fb-gfkh4\" (UID: \"f65fc54c-f0ca-40dc-a8bc-9084483d2d11\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678233 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/27733066-fbcd-474e-9af2-d2bd168533c4-registration-dir\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678257 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk47x\" (UniqueName: \"kubernetes.io/projected/f06c4ffb-6d3f-4189-8e4e-4d5233c41faa-kube-api-access-tk47x\") pod \"service-ca-operator-777779d784-j2hql\" (UID: \"f06c4ffb-6d3f-4189-8e4e-4d5233c41faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-j2hql" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678277 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4z9d\" (UniqueName: \"kubernetes.io/projected/6b5dec00-19df-4fb7-883e-6941ebe98d0a-kube-api-access-l4z9d\") pod \"machine-config-server-nngbv\" (UID: \"6b5dec00-19df-4fb7-883e-6941ebe98d0a\") " pod="openshift-machine-config-operator/machine-config-server-nngbv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678326 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f65fc54c-f0ca-40dc-a8bc-9084483d2d11-srv-cert\") pod \"olm-operator-6b444d44fb-gfkh4\" (UID: \"f65fc54c-f0ca-40dc-a8bc-9084483d2d11\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678349 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vr8lx\" (UniqueName: \"kubernetes.io/projected/f6f43404-a6e1-49ee-bbef-9e027779cef5-kube-api-access-vr8lx\") pod \"ingress-operator-5b745b69d9-rpnhr\" (UID: \"f6f43404-a6e1-49ee-bbef-9e027779cef5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678373 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r24l7\" (UniqueName: \"kubernetes.io/projected/d2b3bc37-36b2-473e-ae10-b0bdc30f2ef2-kube-api-access-r24l7\") pod \"migrator-59844c95c7-r2t78\" (UID: \"d2b3bc37-36b2-473e-ae10-b0bdc30f2ef2\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2t78" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678400 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppmbk\" (UniqueName: \"kubernetes.io/projected/5cc40c37-833b-439a-bfd3-b6aa63b8c033-kube-api-access-ppmbk\") pod \"machine-approver-56656f9798-tvwsv\" (UID: \"5cc40c37-833b-439a-bfd3-b6aa63b8c033\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678416 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f6f43404-a6e1-49ee-bbef-9e027779cef5-trusted-ca\") pod \"ingress-operator-5b745b69d9-rpnhr\" (UID: \"f6f43404-a6e1-49ee-bbef-9e027779cef5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678448 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/27733066-fbcd-474e-9af2-d2bd168533c4-socket-dir\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678464 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlpnb\" (UniqueName: \"kubernetes.io/projected/27733066-fbcd-474e-9af2-d2bd168533c4-kube-api-access-dlpnb\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678487 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb4q6\" (UniqueName: \"kubernetes.io/projected/56c2d16b-fba2-4c54-95fb-71e7676b4445-kube-api-access-wb4q6\") pod \"service-ca-9c57cc56f-k9rh4\" (UID: \"56c2d16b-fba2-4c54-95fb-71e7676b4445\") " pod="openshift-service-ca/service-ca-9c57cc56f-k9rh4" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678505 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/27733066-fbcd-474e-9af2-d2bd168533c4-plugins-dir\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678534 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5cc40c37-833b-439a-bfd3-b6aa63b8c033-machine-approver-tls\") pod \"machine-approver-56656f9798-tvwsv\" (UID: \"5cc40c37-833b-439a-bfd3-b6aa63b8c033\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678550 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f1fce81-bf5b-48df-a92a-e95e82b5ef54-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6hsl2\" (UID: \"9f1fce81-bf5b-48df-a92a-e95e82b5ef54\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678565 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fab8c56-bd27-485d-8216-c219effe9dcc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp64k\" (UID: \"2fab8c56-bd27-485d-8216-c219effe9dcc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp64k" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678765 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/27733066-fbcd-474e-9af2-d2bd168533c4-socket-dir\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678819 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/27733066-fbcd-474e-9af2-d2bd168533c4-plugins-dir\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.678894 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/27733066-fbcd-474e-9af2-d2bd168533c4-registration-dir\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.679147 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fab8c56-bd27-485d-8216-c219effe9dcc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp64k\" (UID: \"2fab8c56-bd27-485d-8216-c219effe9dcc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp64k" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.679583 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f6f43404-a6e1-49ee-bbef-9e027779cef5-trusted-ca\") pod \"ingress-operator-5b745b69d9-rpnhr\" (UID: \"f6f43404-a6e1-49ee-bbef-9e027779cef5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.681917 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f6f43404-a6e1-49ee-bbef-9e027779cef5-metrics-tls\") pod \"ingress-operator-5b745b69d9-rpnhr\" (UID: \"f6f43404-a6e1-49ee-bbef-9e027779cef5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.682857 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fab8c56-bd27-485d-8216-c219effe9dcc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp64k\" (UID: \"2fab8c56-bd27-485d-8216-c219effe9dcc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp64k" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.692443 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.716815 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.738600 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.755823 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.759014 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8de762a2-b5b9-4e05-aedc-9ba4d05ee95f-config\") pod \"kube-controller-manager-operator-78b949d7b-8h4wk\" (UID: \"8de762a2-b5b9-4e05-aedc-9ba4d05ee95f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8h4wk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.772901 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.793403 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.818818 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.832606 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.852381 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.872790 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.890734 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8de762a2-b5b9-4e05-aedc-9ba4d05ee95f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8h4wk\" (UID: \"8de762a2-b5b9-4e05-aedc-9ba4d05ee95f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8h4wk" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.893557 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.912991 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.932160 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.954014 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.972742 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 13:58:19 crc kubenswrapper[4706]: I1206 13:58:19.993654 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.013131 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.032872 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.051761 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.072553 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.092515 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.113033 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.133154 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.153604 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.172792 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.192824 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.214470 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.233301 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.253296 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.263415 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f65fc54c-f0ca-40dc-a8bc-9084483d2d11-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gfkh4\" (UID: \"f65fc54c-f0ca-40dc-a8bc-9084483d2d11\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.300610 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjw7r\" (UniqueName: \"kubernetes.io/projected/4c02a475-2aec-4c9c-ba4d-f20059c09682-kube-api-access-hjw7r\") pod \"machine-api-operator-5694c8668f-9hxp8\" (UID: \"4c02a475-2aec-4c9c-ba4d-f20059c09682\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.323068 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8k5g\" (UniqueName: \"kubernetes.io/projected/475504b3-7830-4226-b00c-3d35eda647fe-kube-api-access-p8k5g\") pod \"controller-manager-879f6c89f-tsvsk\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.333296 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqttw\" (UniqueName: \"kubernetes.io/projected/1a9eee3d-8221-46cf-b949-bd7e854e12ab-kube-api-access-mqttw\") pod \"openshift-apiserver-operator-796bbdcf4f-qgdc9\" (UID: \"1a9eee3d-8221-46cf-b949-bd7e854e12ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qgdc9" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.360406 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wknkv\" (UniqueName: \"kubernetes.io/projected/0136a364-d7ad-445a-9226-5b064e4b4527-kube-api-access-wknkv\") pod \"apiserver-7bbb656c7d-dfd8q\" (UID: \"0136a364-d7ad-445a-9226-5b064e4b4527\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.372902 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.392355 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.412360 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.432383 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.444572 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f65fc54c-f0ca-40dc-a8bc-9084483d2d11-srv-cert\") pod \"olm-operator-6b444d44fb-gfkh4\" (UID: \"f65fc54c-f0ca-40dc-a8bc-9084483d2d11\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.452979 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.472414 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.493851 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.510673 4706 request.go:700] Waited for 1.004404554s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-cluster-machine-approver/secrets?fieldSelector=metadata.name%3Dmachine-approver-sa-dockercfg-nl2j4&limit=500&resourceVersion=0 Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.513651 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.531958 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.532578 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.546433 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cc40c37-833b-439a-bfd3-b6aa63b8c033-config\") pod \"machine-approver-56656f9798-tvwsv\" (UID: \"5cc40c37-833b-439a-bfd3-b6aa63b8c033\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.546896 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.555559 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.559290 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5cc40c37-833b-439a-bfd3-b6aa63b8c033-auth-proxy-config\") pod \"machine-approver-56656f9798-tvwsv\" (UID: \"5cc40c37-833b-439a-bfd3-b6aa63b8c033\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.566784 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.573655 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.594268 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.604482 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5cc40c37-833b-439a-bfd3-b6aa63b8c033-machine-approver-tls\") pod \"machine-approver-56656f9798-tvwsv\" (UID: \"5cc40c37-833b-439a-bfd3-b6aa63b8c033\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.613100 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.619195 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qgdc9" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.633811 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.653673 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.673113 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678410 4706 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678499 4706 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678508 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b5dec00-19df-4fb7-883e-6941ebe98d0a-node-bootstrap-token podName:6b5dec00-19df-4fb7-883e-6941ebe98d0a nodeName:}" failed. No retries permitted until 2025-12-06 13:58:21.178484645 +0000 UTC m=+144.014277502 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/6b5dec00-19df-4fb7-883e-6941ebe98d0a-node-bootstrap-token") pod "machine-config-server-nngbv" (UID: "6b5dec00-19df-4fb7-883e-6941ebe98d0a") : failed to sync secret cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678516 4706 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678578 4706 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678605 4706 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678560 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/56c2d16b-fba2-4c54-95fb-71e7676b4445-signing-cabundle podName:56c2d16b-fba2-4c54-95fb-71e7676b4445 nodeName:}" failed. No retries permitted until 2025-12-06 13:58:21.178540446 +0000 UTC m=+144.014333303 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/56c2d16b-fba2-4c54-95fb-71e7676b4445-signing-cabundle") pod "service-ca-9c57cc56f-k9rh4" (UID: "56c2d16b-fba2-4c54-95fb-71e7676b4445") : failed to sync configmap cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678587 4706 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678653 4706 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678672 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f06c4ffb-6d3f-4189-8e4e-4d5233c41faa-serving-cert podName:f06c4ffb-6d3f-4189-8e4e-4d5233c41faa nodeName:}" failed. No retries permitted until 2025-12-06 13:58:21.178641639 +0000 UTC m=+144.014434536 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/f06c4ffb-6d3f-4189-8e4e-4d5233c41faa-serving-cert") pod "service-ca-operator-777779d784-j2hql" (UID: "f06c4ffb-6d3f-4189-8e4e-4d5233c41faa") : failed to sync secret cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678714 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/56c2d16b-fba2-4c54-95fb-71e7676b4445-signing-key podName:56c2d16b-fba2-4c54-95fb-71e7676b4445 nodeName:}" failed. No retries permitted until 2025-12-06 13:58:21.17869661 +0000 UTC m=+144.014489617 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/56c2d16b-fba2-4c54-95fb-71e7676b4445-signing-key") pod "service-ca-9c57cc56f-k9rh4" (UID: "56c2d16b-fba2-4c54-95fb-71e7676b4445") : failed to sync secret cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678743 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-trusted-ca podName:c2e4f1d6-67ff-4990-bcbf-13ad23dc511b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:21.178731321 +0000 UTC m=+144.014524218 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-trusted-ca") pod "marketplace-operator-79b997595-cvtpw" (UID: "c2e4f1d6-67ff-4990-bcbf-13ad23dc511b") : failed to sync configmap cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678791 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-operator-metrics podName:c2e4f1d6-67ff-4990-bcbf-13ad23dc511b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:21.178754322 +0000 UTC m=+144.014547219 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-operator-metrics") pod "marketplace-operator-79b997595-cvtpw" (UID: "c2e4f1d6-67ff-4990-bcbf-13ad23dc511b") : failed to sync secret cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678822 4706 configmap.go:193] Couldn't get configMap openshift-kube-storage-version-migrator-operator/config: failed to sync configmap cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678825 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b5dec00-19df-4fb7-883e-6941ebe98d0a-certs podName:6b5dec00-19df-4fb7-883e-6941ebe98d0a nodeName:}" failed. No retries permitted until 2025-12-06 13:58:21.178808683 +0000 UTC m=+144.014601580 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/6b5dec00-19df-4fb7-883e-6941ebe98d0a-certs") pod "machine-config-server-nngbv" (UID: "6b5dec00-19df-4fb7-883e-6941ebe98d0a") : failed to sync secret cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678898 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9f1fce81-bf5b-48df-a92a-e95e82b5ef54-config podName:9f1fce81-bf5b-48df-a92a-e95e82b5ef54 nodeName:}" failed. No retries permitted until 2025-12-06 13:58:21.178884095 +0000 UTC m=+144.014677072 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/9f1fce81-bf5b-48df-a92a-e95e82b5ef54-config") pod "kube-storage-version-migrator-operator-b67b599dd-6hsl2" (UID: "9f1fce81-bf5b-48df-a92a-e95e82b5ef54") : failed to sync configmap cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678904 4706 secret.go:188] Couldn't get secret openshift-kube-storage-version-migrator-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678869 4706 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.678987 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9f1fce81-bf5b-48df-a92a-e95e82b5ef54-serving-cert podName:9f1fce81-bf5b-48df-a92a-e95e82b5ef54 nodeName:}" failed. No retries permitted until 2025-12-06 13:58:21.178970268 +0000 UTC m=+144.014763165 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/9f1fce81-bf5b-48df-a92a-e95e82b5ef54-serving-cert") pod "kube-storage-version-migrator-operator-b67b599dd-6hsl2" (UID: "9f1fce81-bf5b-48df-a92a-e95e82b5ef54") : failed to sync secret cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: E1206 13:58:20.679053 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f06c4ffb-6d3f-4189-8e4e-4d5233c41faa-config podName:f06c4ffb-6d3f-4189-8e4e-4d5233c41faa nodeName:}" failed. No retries permitted until 2025-12-06 13:58:21.178999948 +0000 UTC m=+144.014792835 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/f06c4ffb-6d3f-4189-8e4e-4d5233c41faa-config") pod "service-ca-operator-777779d784-j2hql" (UID: "f06c4ffb-6d3f-4189-8e4e-4d5233c41faa") : failed to sync configmap cache: timed out waiting for the condition Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.692613 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.715035 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.732542 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.752954 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.772660 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.791936 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.806399 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q"] Dec 06 13:58:20 crc kubenswrapper[4706]: W1206 13:58:20.820208 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0136a364_d7ad_445a_9226_5b064e4b4527.slice/crio-d70fdd4649013b6fa9627276ea44cb6c9f1cf6ad1daa7391c5d9422f0deef785 WatchSource:0}: Error finding container d70fdd4649013b6fa9627276ea44cb6c9f1cf6ad1daa7391c5d9422f0deef785: Status 404 returned error can't find the container with id d70fdd4649013b6fa9627276ea44cb6c9f1cf6ad1daa7391c5d9422f0deef785 Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.820532 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.833922 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.852966 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.861586 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qgdc9"] Dec 06 13:58:20 crc kubenswrapper[4706]: W1206 13:58:20.869457 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a9eee3d_8221_46cf_b949_bd7e854e12ab.slice/crio-92604a15a141e429115901ad7b30549cf67b865b70d15e1515fed6310d717328 WatchSource:0}: Error finding container 92604a15a141e429115901ad7b30549cf67b865b70d15e1515fed6310d717328: Status 404 returned error can't find the container with id 92604a15a141e429115901ad7b30549cf67b865b70d15e1515fed6310d717328 Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.872735 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.892651 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.912779 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.932229 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.953600 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.973595 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 13:58:20 crc kubenswrapper[4706]: I1206 13:58:20.992699 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.013631 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.031683 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.034529 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9hxp8"] Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.041413 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tsvsk"] Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.052691 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.072638 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.092867 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 13:58:21 crc kubenswrapper[4706]: W1206 13:58:21.099050 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod475504b3_7830_4226_b00c_3d35eda647fe.slice/crio-b0aaca3181268ca5d05f5decc481795b9f531d9bd6ebe9f9384b93ea16a3f08a WatchSource:0}: Error finding container b0aaca3181268ca5d05f5decc481795b9f531d9bd6ebe9f9384b93ea16a3f08a: Status 404 returned error can't find the container with id b0aaca3181268ca5d05f5decc481795b9f531d9bd6ebe9f9384b93ea16a3f08a Dec 06 13:58:21 crc kubenswrapper[4706]: W1206 13:58:21.100631 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c02a475_2aec_4c9c_ba4d_f20059c09682.slice/crio-87316f0f2b5cbc97aa99342d1891783c5ca47e75d8eccac2be68a1c8ad959e12 WatchSource:0}: Error finding container 87316f0f2b5cbc97aa99342d1891783c5ca47e75d8eccac2be68a1c8ad959e12: Status 404 returned error can't find the container with id 87316f0f2b5cbc97aa99342d1891783c5ca47e75d8eccac2be68a1c8ad959e12 Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.113461 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.133558 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.152764 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.174501 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.192876 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.212097 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cvtpw\" (UID: \"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.212164 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/56c2d16b-fba2-4c54-95fb-71e7676b4445-signing-key\") pod \"service-ca-9c57cc56f-k9rh4\" (UID: \"56c2d16b-fba2-4c54-95fb-71e7676b4445\") " pod="openshift-service-ca/service-ca-9c57cc56f-k9rh4" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.212200 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f1fce81-bf5b-48df-a92a-e95e82b5ef54-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6hsl2\" (UID: \"9f1fce81-bf5b-48df-a92a-e95e82b5ef54\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.212265 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6b5dec00-19df-4fb7-883e-6941ebe98d0a-node-bootstrap-token\") pod \"machine-config-server-nngbv\" (UID: \"6b5dec00-19df-4fb7-883e-6941ebe98d0a\") " pod="openshift-machine-config-operator/machine-config-server-nngbv" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.212299 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f06c4ffb-6d3f-4189-8e4e-4d5233c41faa-serving-cert\") pod \"service-ca-operator-777779d784-j2hql\" (UID: \"f06c4ffb-6d3f-4189-8e4e-4d5233c41faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-j2hql" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.212394 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f06c4ffb-6d3f-4189-8e4e-4d5233c41faa-config\") pod \"service-ca-operator-777779d784-j2hql\" (UID: \"f06c4ffb-6d3f-4189-8e4e-4d5233c41faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-j2hql" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.212432 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cvtpw\" (UID: \"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.212680 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f1fce81-bf5b-48df-a92a-e95e82b5ef54-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6hsl2\" (UID: \"9f1fce81-bf5b-48df-a92a-e95e82b5ef54\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.212729 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.212732 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/56c2d16b-fba2-4c54-95fb-71e7676b4445-signing-cabundle\") pod \"service-ca-9c57cc56f-k9rh4\" (UID: \"56c2d16b-fba2-4c54-95fb-71e7676b4445\") " pod="openshift-service-ca/service-ca-9c57cc56f-k9rh4" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.212971 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6b5dec00-19df-4fb7-883e-6941ebe98d0a-certs\") pod \"machine-config-server-nngbv\" (UID: \"6b5dec00-19df-4fb7-883e-6941ebe98d0a\") " pod="openshift-machine-config-operator/machine-config-server-nngbv" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.213998 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/56c2d16b-fba2-4c54-95fb-71e7676b4445-signing-cabundle\") pod \"service-ca-9c57cc56f-k9rh4\" (UID: \"56c2d16b-fba2-4c54-95fb-71e7676b4445\") " pod="openshift-service-ca/service-ca-9c57cc56f-k9rh4" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.213998 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f06c4ffb-6d3f-4189-8e4e-4d5233c41faa-config\") pod \"service-ca-operator-777779d784-j2hql\" (UID: \"f06c4ffb-6d3f-4189-8e4e-4d5233c41faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-j2hql" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.214609 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f1fce81-bf5b-48df-a92a-e95e82b5ef54-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6hsl2\" (UID: \"9f1fce81-bf5b-48df-a92a-e95e82b5ef54\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.215778 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cvtpw\" (UID: \"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.218133 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f1fce81-bf5b-48df-a92a-e95e82b5ef54-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6hsl2\" (UID: \"9f1fce81-bf5b-48df-a92a-e95e82b5ef54\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.218183 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f06c4ffb-6d3f-4189-8e4e-4d5233c41faa-serving-cert\") pod \"service-ca-operator-777779d784-j2hql\" (UID: \"f06c4ffb-6d3f-4189-8e4e-4d5233c41faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-j2hql" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.219484 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cvtpw\" (UID: \"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.219726 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/56c2d16b-fba2-4c54-95fb-71e7676b4445-signing-key\") pod \"service-ca-9c57cc56f-k9rh4\" (UID: \"56c2d16b-fba2-4c54-95fb-71e7676b4445\") " pod="openshift-service-ca/service-ca-9c57cc56f-k9rh4" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.235551 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.254109 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.268373 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6b5dec00-19df-4fb7-883e-6941ebe98d0a-node-bootstrap-token\") pod \"machine-config-server-nngbv\" (UID: \"6b5dec00-19df-4fb7-883e-6941ebe98d0a\") " pod="openshift-machine-config-operator/machine-config-server-nngbv" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.272632 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.276756 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6b5dec00-19df-4fb7-883e-6941ebe98d0a-certs\") pod \"machine-config-server-nngbv\" (UID: \"6b5dec00-19df-4fb7-883e-6941ebe98d0a\") " pod="openshift-machine-config-operator/machine-config-server-nngbv" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.293473 4706 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.313243 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.316289 4706 generic.go:334] "Generic (PLEG): container finished" podID="0136a364-d7ad-445a-9226-5b064e4b4527" containerID="40add6f61b38d42cdef84afbb9848bcebbe9e1e4bc9c31db06b88e9498af68d4" exitCode=0 Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.316367 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" event={"ID":"0136a364-d7ad-445a-9226-5b064e4b4527","Type":"ContainerDied","Data":"40add6f61b38d42cdef84afbb9848bcebbe9e1e4bc9c31db06b88e9498af68d4"} Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.316391 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" event={"ID":"0136a364-d7ad-445a-9226-5b064e4b4527","Type":"ContainerStarted","Data":"d70fdd4649013b6fa9627276ea44cb6c9f1cf6ad1daa7391c5d9422f0deef785"} Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.319864 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" event={"ID":"4c02a475-2aec-4c9c-ba4d-f20059c09682","Type":"ContainerStarted","Data":"efbffba59c2ee937631b6797e234cfe80f282a77cf07f91c3266fbf630f50547"} Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.319912 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" event={"ID":"4c02a475-2aec-4c9c-ba4d-f20059c09682","Type":"ContainerStarted","Data":"9495671726f645653b4553175f873aa69d71ba454e2e37638bc666906139ab22"} Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.319938 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" event={"ID":"4c02a475-2aec-4c9c-ba4d-f20059c09682","Type":"ContainerStarted","Data":"87316f0f2b5cbc97aa99342d1891783c5ca47e75d8eccac2be68a1c8ad959e12"} Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.321294 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" event={"ID":"475504b3-7830-4226-b00c-3d35eda647fe","Type":"ContainerStarted","Data":"84d8095aad61a3e7266019b5357582efdd2a1549358ef81da95f9ea47ecf7cee"} Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.321342 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" event={"ID":"475504b3-7830-4226-b00c-3d35eda647fe","Type":"ContainerStarted","Data":"b0aaca3181268ca5d05f5decc481795b9f531d9bd6ebe9f9384b93ea16a3f08a"} Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.321528 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.322789 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qgdc9" event={"ID":"1a9eee3d-8221-46cf-b949-bd7e854e12ab","Type":"ContainerStarted","Data":"86e8a639d0e35d48efb4ec0b2ef796b4ef29663b0ee356d7f8323fe4c6bbb2b5"} Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.322817 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qgdc9" event={"ID":"1a9eee3d-8221-46cf-b949-bd7e854e12ab","Type":"ContainerStarted","Data":"92604a15a141e429115901ad7b30549cf67b865b70d15e1515fed6310d717328"} Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.329819 4706 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-tsvsk container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.329867 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" podUID="475504b3-7830-4226-b00c-3d35eda647fe" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.332725 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.372811 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.394336 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.412700 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.434542 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.453798 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.484599 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.491866 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.510719 4706 request.go:700] Waited for 1.940990083s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console/serviceaccounts/default/token Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.529656 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7nbw\" (UniqueName: \"kubernetes.io/projected/621a48da-c965-4d74-abd0-62e119fd81c8-kube-api-access-v7nbw\") pod \"downloads-7954f5f757-9jn7l\" (UID: \"621a48da-c965-4d74-abd0-62e119fd81c8\") " pod="openshift-console/downloads-7954f5f757-9jn7l" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.546217 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bhm9\" (UniqueName: \"kubernetes.io/projected/caf8bd48-353a-428e-ba7a-1a7c40f19b23-kube-api-access-6bhm9\") pod \"openshift-config-operator-7777fb866f-2wmlg\" (UID: \"caf8bd48-353a-428e-ba7a-1a7c40f19b23\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.574084 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9kn2\" (UniqueName: \"kubernetes.io/projected/bc8af660-99b4-4d61-bc9c-4580f285d6c6-kube-api-access-f9kn2\") pod \"openshift-controller-manager-operator-756b6f6bc6-2nk8h\" (UID: \"bc8af660-99b4-4d61-bc9c-4580f285d6c6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2nk8h" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.592231 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w54rd\" (UniqueName: \"kubernetes.io/projected/77b4e41c-830d-48c2-8d00-97b33c7e8a29-kube-api-access-w54rd\") pod \"control-plane-machine-set-operator-78cbb6b69f-p85f8\" (UID: \"77b4e41c-830d-48c2-8d00-97b33c7e8a29\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p85f8" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.599427 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.606216 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5hnr\" (UniqueName: \"kubernetes.io/projected/b38e2a11-98f9-47c3-a711-106d6aa308de-kube-api-access-k5hnr\") pod \"cluster-samples-operator-665b6dd947-595n2\" (UID: \"b38e2a11-98f9-47c3-a711-106d6aa308de\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-595n2" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.611545 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9jn7l" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.618349 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p85f8" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.623126 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2nk8h" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.632723 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx6lf\" (UniqueName: \"kubernetes.io/projected/fe23e3f5-57af-4fd2-a8bd-9364bbe39776-kube-api-access-vx6lf\") pod \"authentication-operator-69f744f599-j4chj\" (UID: \"fe23e3f5-57af-4fd2-a8bd-9364bbe39776\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.662235 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8xz9\" (UniqueName: \"kubernetes.io/projected/b801f158-2df3-44c7-8bfe-5caa8467e29d-kube-api-access-d8xz9\") pod \"route-controller-manager-6576b87f9c-dmfw8\" (UID: \"b801f158-2df3-44c7-8bfe-5caa8467e29d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.691193 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqhbg\" (UniqueName: \"kubernetes.io/projected/b9f16df2-5700-43c2-9488-78c3a9d8bcd6-kube-api-access-kqhbg\") pod \"router-default-5444994796-r94jx\" (UID: \"b9f16df2-5700-43c2-9488-78c3a9d8bcd6\") " pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.718147 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhks2\" (UniqueName: \"kubernetes.io/projected/0acf2c47-e9d7-4759-aa80-9f4ca8488899-kube-api-access-mhks2\") pod \"apiserver-76f77b778f-p8xpt\" (UID: \"0acf2c47-e9d7-4759-aa80-9f4ca8488899\") " pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.718902 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f6f43404-a6e1-49ee-bbef-9e027779cef5-bound-sa-token\") pod \"ingress-operator-5b745b69d9-rpnhr\" (UID: \"f6f43404-a6e1-49ee-bbef-9e027779cef5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.741783 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8de762a2-b5b9-4e05-aedc-9ba4d05ee95f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8h4wk\" (UID: \"8de762a2-b5b9-4e05-aedc-9ba4d05ee95f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8h4wk" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.765633 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzjxf\" (UniqueName: \"kubernetes.io/projected/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-kube-api-access-tzjxf\") pod \"marketplace-operator-79b997595-cvtpw\" (UID: \"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.769787 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk47x\" (UniqueName: \"kubernetes.io/projected/f06c4ffb-6d3f-4189-8e4e-4d5233c41faa-kube-api-access-tk47x\") pod \"service-ca-operator-777779d784-j2hql\" (UID: \"f06c4ffb-6d3f-4189-8e4e-4d5233c41faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-j2hql" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.785323 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.791302 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2fab8c56-bd27-485d-8216-c219effe9dcc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp64k\" (UID: \"2fab8c56-bd27-485d-8216-c219effe9dcc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp64k" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.803979 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-j2hql" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.811007 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vr8lx\" (UniqueName: \"kubernetes.io/projected/f6f43404-a6e1-49ee-bbef-9e027779cef5-kube-api-access-vr8lx\") pod \"ingress-operator-5b745b69d9-rpnhr\" (UID: \"f6f43404-a6e1-49ee-bbef-9e027779cef5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.827760 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.840774 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz264\" (UniqueName: \"kubernetes.io/projected/f65fc54c-f0ca-40dc-a8bc-9084483d2d11-kube-api-access-lz264\") pod \"olm-operator-6b444d44fb-gfkh4\" (UID: \"f65fc54c-f0ca-40dc-a8bc-9084483d2d11\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.845606 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.854973 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4z9d\" (UniqueName: \"kubernetes.io/projected/6b5dec00-19df-4fb7-883e-6941ebe98d0a-kube-api-access-l4z9d\") pod \"machine-config-server-nngbv\" (UID: \"6b5dec00-19df-4fb7-883e-6941ebe98d0a\") " pod="openshift-machine-config-operator/machine-config-server-nngbv" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.856247 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.867592 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-595n2" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.882914 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppmbk\" (UniqueName: \"kubernetes.io/projected/5cc40c37-833b-439a-bfd3-b6aa63b8c033-kube-api-access-ppmbk\") pod \"machine-approver-56656f9798-tvwsv\" (UID: \"5cc40c37-833b-439a-bfd3-b6aa63b8c033\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.896485 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb4q6\" (UniqueName: \"kubernetes.io/projected/56c2d16b-fba2-4c54-95fb-71e7676b4445-kube-api-access-wb4q6\") pod \"service-ca-9c57cc56f-k9rh4\" (UID: \"56c2d16b-fba2-4c54-95fb-71e7676b4445\") " pod="openshift-service-ca/service-ca-9c57cc56f-k9rh4" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.909662 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlpnb\" (UniqueName: \"kubernetes.io/projected/27733066-fbcd-474e-9af2-d2bd168533c4-kube-api-access-dlpnb\") pod \"csi-hostpathplugin-7xn8m\" (UID: \"27733066-fbcd-474e-9af2-d2bd168533c4\") " pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.929044 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.930083 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r24l7\" (UniqueName: \"kubernetes.io/projected/d2b3bc37-36b2-473e-ae10-b0bdc30f2ef2-kube-api-access-r24l7\") pod \"migrator-59844c95c7-r2t78\" (UID: \"d2b3bc37-36b2-473e-ae10-b0bdc30f2ef2\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2t78" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.938752 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p85f8"] Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.942056 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp64k" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.948608 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.963998 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v26w\" (UniqueName: \"kubernetes.io/projected/9f1fce81-bf5b-48df-a92a-e95e82b5ef54-kube-api-access-8v26w\") pod \"kube-storage-version-migrator-operator-b67b599dd-6hsl2\" (UID: \"9f1fce81-bf5b-48df-a92a-e95e82b5ef54\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2" Dec 06 13:58:21 crc kubenswrapper[4706]: I1206 13:58:21.969555 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9jn7l"] Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.010774 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8h4wk" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.029928 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz285\" (UniqueName: \"kubernetes.io/projected/806ad18b-e408-4679-9a37-b6efd48dd4d8-kube-api-access-bz285\") pod \"package-server-manager-789f6589d5-dp75h\" (UID: \"806ad18b-e408-4679-9a37-b6efd48dd4d8\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dp75h" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.029981 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7898190a-5957-4772-abbb-1bfbf16f4eb5-audit-dir\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.030009 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-registry-tls\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.030027 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d493be9f-6d12-4d4c-a04b-da608ff57eaa-serving-cert\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.030041 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msqjg\" (UniqueName: \"kubernetes.io/projected/1a8d709a-565d-4213-ae67-17440c2d7678-kube-api-access-msqjg\") pod \"multus-admission-controller-857f4d67dd-zx8fv\" (UID: \"1a8d709a-565d-4213-ae67-17440c2d7678\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zx8fv" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.030057 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d493be9f-6d12-4d4c-a04b-da608ff57eaa-etcd-client\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.030090 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d0d1509c-c26c-4066-8dab-23d0205d4211-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2qmsq\" (UID: \"d0d1509c-c26c-4066-8dab-23d0205d4211\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.030109 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.030124 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6b88\" (UniqueName: \"kubernetes.io/projected/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-kube-api-access-g6b88\") pod \"collect-profiles-29417145-v956j\" (UID: \"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.030610 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d493be9f-6d12-4d4c-a04b-da608ff57eaa-config\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.030657 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnvvv\" (UniqueName: \"kubernetes.io/projected/1ae74e67-c9db-4ba6-87e5-0fcb923244b3-kube-api-access-cnvvv\") pod \"machine-config-controller-84d6567774-sx2jx\" (UID: \"1ae74e67-c9db-4ba6-87e5-0fcb923244b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.030717 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1ae74e67-c9db-4ba6-87e5-0fcb923244b3-proxy-tls\") pod \"machine-config-controller-84d6567774-sx2jx\" (UID: \"1ae74e67-c9db-4ba6-87e5-0fcb923244b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.030741 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ed999748-8e6d-4402-81e3-1f483b8440e8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.030758 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.030851 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/532ce376-db1f-4770-a4cd-a30318242daf-console-serving-cert\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.030908 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzxhl\" (UniqueName: \"kubernetes.io/projected/aca05689-5cad-4e13-b513-1ef9de804ba9-kube-api-access-mzxhl\") pod \"catalog-operator-68c6474976-mfhhs\" (UID: \"aca05689-5cad-4e13-b513-1ef9de804ba9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.030944 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d0d1509c-c26c-4066-8dab-23d0205d4211-images\") pod \"machine-config-operator-74547568cd-2qmsq\" (UID: \"d0d1509c-c26c-4066-8dab-23d0205d4211\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.030980 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed999748-8e6d-4402-81e3-1f483b8440e8-trusted-ca\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.031000 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d493be9f-6d12-4d4c-a04b-da608ff57eaa-etcd-service-ca\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.031605 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0b36426-e051-4864-b98e-b1ecf89914a4-config\") pod \"console-operator-58897d9998-t2255\" (UID: \"a0b36426-e051-4864-b98e-b1ecf89914a4\") " pod="openshift-console-operator/console-operator-58897d9998-t2255" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.031636 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65a55c43-6a12-46dc-b181-8241ceab5d12-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-97565\" (UID: \"65a55c43-6a12-46dc-b181-8241ceab5d12\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-97565" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.031657 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-secret-volume\") pod \"collect-profiles-29417145-v956j\" (UID: \"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.031693 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-bound-sa-token\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.031720 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/532ce376-db1f-4770-a4cd-a30318242daf-trusted-ca-bundle\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.031759 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1a8d709a-565d-4213-ae67-17440c2d7678-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-zx8fv\" (UID: \"1a8d709a-565d-4213-ae67-17440c2d7678\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zx8fv" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.031781 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.031807 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0b36426-e051-4864-b98e-b1ecf89914a4-serving-cert\") pod \"console-operator-58897d9998-t2255\" (UID: \"a0b36426-e051-4864-b98e-b1ecf89914a4\") " pod="openshift-console-operator/console-operator-58897d9998-t2255" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.031828 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-bkh56\" (UID: \"1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.031848 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/532ce376-db1f-4770-a4cd-a30318242daf-console-config\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.031866 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/532ce376-db1f-4770-a4cd-a30318242daf-oauth-serving-cert\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.031909 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlrfv\" (UniqueName: \"kubernetes.io/projected/09f7b3f2-ef36-4a21-970a-3a19b7a6a726-kube-api-access-dlrfv\") pod \"packageserver-d55dfcdfc-d5zwg\" (UID: \"09f7b3f2-ef36-4a21-970a-3a19b7a6a726\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.031962 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1ae74e67-c9db-4ba6-87e5-0fcb923244b3-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sx2jx\" (UID: \"1ae74e67-c9db-4ba6-87e5-0fcb923244b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.031992 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032017 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d0d1509c-c26c-4066-8dab-23d0205d4211-proxy-tls\") pod \"machine-config-operator-74547568cd-2qmsq\" (UID: \"d0d1509c-c26c-4066-8dab-23d0205d4211\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032037 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-audit-policies\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032069 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghxlc\" (UniqueName: \"kubernetes.io/projected/1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90-kube-api-access-ghxlc\") pod \"cluster-image-registry-operator-dc59b4c8b-bkh56\" (UID: \"1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032102 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/65a55c43-6a12-46dc-b181-8241ceab5d12-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-97565\" (UID: \"65a55c43-6a12-46dc-b181-8241ceab5d12\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-97565" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032126 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032157 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032195 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032220 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwm75\" (UniqueName: \"kubernetes.io/projected/532ce376-db1f-4770-a4cd-a30318242daf-kube-api-access-zwm75\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032242 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032280 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ed999748-8e6d-4402-81e3-1f483b8440e8-registry-certificates\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032334 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/09f7b3f2-ef36-4a21-970a-3a19b7a6a726-webhook-cert\") pod \"packageserver-d55dfcdfc-d5zwg\" (UID: \"09f7b3f2-ef36-4a21-970a-3a19b7a6a726\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032373 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxfnb\" (UniqueName: \"kubernetes.io/projected/d0d1509c-c26c-4066-8dab-23d0205d4211-kube-api-access-bxfnb\") pod \"machine-config-operator-74547568cd-2qmsq\" (UID: \"d0d1509c-c26c-4066-8dab-23d0205d4211\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032396 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032422 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/09f7b3f2-ef36-4a21-970a-3a19b7a6a726-apiservice-cert\") pod \"packageserver-d55dfcdfc-d5zwg\" (UID: \"09f7b3f2-ef36-4a21-970a-3a19b7a6a726\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032457 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2daa96a6-fc64-4fd3-ba1a-d4ae6bbf97b0-metrics-tls\") pod \"dns-operator-744455d44c-cl6js\" (UID: \"2daa96a6-fc64-4fd3-ba1a-d4ae6bbf97b0\") " pod="openshift-dns-operator/dns-operator-744455d44c-cl6js" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032477 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/09f7b3f2-ef36-4a21-970a-3a19b7a6a726-tmpfs\") pod \"packageserver-d55dfcdfc-d5zwg\" (UID: \"09f7b3f2-ef36-4a21-970a-3a19b7a6a726\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032501 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trqvg\" (UniqueName: \"kubernetes.io/projected/2daa96a6-fc64-4fd3-ba1a-d4ae6bbf97b0-kube-api-access-trqvg\") pod \"dns-operator-744455d44c-cl6js\" (UID: \"2daa96a6-fc64-4fd3-ba1a-d4ae6bbf97b0\") " pod="openshift-dns-operator/dns-operator-744455d44c-cl6js" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032527 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d493be9f-6d12-4d4c-a04b-da608ff57eaa-etcd-ca\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032606 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/532ce376-db1f-4770-a4cd-a30318242daf-console-oauth-config\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032639 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032662 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-config-volume\") pod \"collect-profiles-29417145-v956j\" (UID: \"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.032686 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.033211 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/532ce376-db1f-4770-a4cd-a30318242daf-service-ca\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.033238 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-bkh56\" (UID: \"1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.033275 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j86cv\" (UniqueName: \"kubernetes.io/projected/a0b36426-e051-4864-b98e-b1ecf89914a4-kube-api-access-j86cv\") pod \"console-operator-58897d9998-t2255\" (UID: \"a0b36426-e051-4864-b98e-b1ecf89914a4\") " pod="openshift-console-operator/console-operator-58897d9998-t2255" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.033294 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/aca05689-5cad-4e13-b513-1ef9de804ba9-profile-collector-cert\") pod \"catalog-operator-68c6474976-mfhhs\" (UID: \"aca05689-5cad-4e13-b513-1ef9de804ba9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.033358 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/aca05689-5cad-4e13-b513-1ef9de804ba9-srv-cert\") pod \"catalog-operator-68c6474976-mfhhs\" (UID: \"aca05689-5cad-4e13-b513-1ef9de804ba9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.033383 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp5w4\" (UniqueName: \"kubernetes.io/projected/d493be9f-6d12-4d4c-a04b-da608ff57eaa-kube-api-access-dp5w4\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.033436 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ed999748-8e6d-4402-81e3-1f483b8440e8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.033483 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l25h5\" (UniqueName: \"kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-kube-api-access-l25h5\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.033503 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65a55c43-6a12-46dc-b181-8241ceab5d12-config\") pod \"kube-apiserver-operator-766d6c64bb-97565\" (UID: \"65a55c43-6a12-46dc-b181-8241ceab5d12\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-97565" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.033524 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/806ad18b-e408-4679-9a37-b6efd48dd4d8-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-dp75h\" (UID: \"806ad18b-e408-4679-9a37-b6efd48dd4d8\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dp75h" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.033539 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-bkh56\" (UID: \"1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56" Dec 06 13:58:22 crc kubenswrapper[4706]: E1206 13:58:22.034544 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:22.534530159 +0000 UTC m=+145.370323106 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.035030 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a0b36426-e051-4864-b98e-b1ecf89914a4-trusted-ca\") pod \"console-operator-58897d9998-t2255\" (UID: \"a0b36426-e051-4864-b98e-b1ecf89914a4\") " pod="openshift-console-operator/console-operator-58897d9998-t2255" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.035115 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.035140 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bc9m\" (UniqueName: \"kubernetes.io/projected/7898190a-5957-4772-abbb-1bfbf16f4eb5-kube-api-access-9bc9m\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.035244 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2t78" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.062979 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.073567 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.078069 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-k9rh4" Dec 06 13:58:22 crc kubenswrapper[4706]: W1206 13:58:22.087967 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77b4e41c_830d_48c2_8d00_97b33c7e8a29.slice/crio-316c6234af61df4305df9aa3438f10bfa5e03b6d9febe56ce1265c4c33fc9489 WatchSource:0}: Error finding container 316c6234af61df4305df9aa3438f10bfa5e03b6d9febe56ce1265c4c33fc9489: Status 404 returned error can't find the container with id 316c6234af61df4305df9aa3438f10bfa5e03b6d9febe56ce1265c4c33fc9489 Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.117377 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.118356 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-nngbv" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140352 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140568 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/532ce376-db1f-4770-a4cd-a30318242daf-console-oauth-config\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140591 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-config-volume\") pod \"collect-profiles-29417145-v956j\" (UID: \"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140610 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140628 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/532ce376-db1f-4770-a4cd-a30318242daf-service-ca\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140659 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j86cv\" (UniqueName: \"kubernetes.io/projected/a0b36426-e051-4864-b98e-b1ecf89914a4-kube-api-access-j86cv\") pod \"console-operator-58897d9998-t2255\" (UID: \"a0b36426-e051-4864-b98e-b1ecf89914a4\") " pod="openshift-console-operator/console-operator-58897d9998-t2255" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140676 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/aca05689-5cad-4e13-b513-1ef9de804ba9-profile-collector-cert\") pod \"catalog-operator-68c6474976-mfhhs\" (UID: \"aca05689-5cad-4e13-b513-1ef9de804ba9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140691 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-bkh56\" (UID: \"1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140714 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/aca05689-5cad-4e13-b513-1ef9de804ba9-srv-cert\") pod \"catalog-operator-68c6474976-mfhhs\" (UID: \"aca05689-5cad-4e13-b513-1ef9de804ba9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140746 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp5w4\" (UniqueName: \"kubernetes.io/projected/d493be9f-6d12-4d4c-a04b-da608ff57eaa-kube-api-access-dp5w4\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140773 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ed999748-8e6d-4402-81e3-1f483b8440e8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140806 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38703ec7-616b-499d-a173-4ea254e11d9a-config-volume\") pod \"dns-default-m9fbd\" (UID: \"38703ec7-616b-499d-a173-4ea254e11d9a\") " pod="openshift-dns/dns-default-m9fbd" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140823 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l25h5\" (UniqueName: \"kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-kube-api-access-l25h5\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140841 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65a55c43-6a12-46dc-b181-8241ceab5d12-config\") pod \"kube-apiserver-operator-766d6c64bb-97565\" (UID: \"65a55c43-6a12-46dc-b181-8241ceab5d12\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-97565" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140860 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-bkh56\" (UID: \"1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140877 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctvsb\" (UniqueName: \"kubernetes.io/projected/407db426-50cd-413d-bd5a-392aed7b1897-kube-api-access-ctvsb\") pod \"ingress-canary-kdfk5\" (UID: \"407db426-50cd-413d-bd5a-392aed7b1897\") " pod="openshift-ingress-canary/ingress-canary-kdfk5" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140903 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/806ad18b-e408-4679-9a37-b6efd48dd4d8-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-dp75h\" (UID: \"806ad18b-e408-4679-9a37-b6efd48dd4d8\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dp75h" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.140925 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a0b36426-e051-4864-b98e-b1ecf89914a4-trusted-ca\") pod \"console-operator-58897d9998-t2255\" (UID: \"a0b36426-e051-4864-b98e-b1ecf89914a4\") " pod="openshift-console-operator/console-operator-58897d9998-t2255" Dec 06 13:58:22 crc kubenswrapper[4706]: E1206 13:58:22.140945 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:22.640925673 +0000 UTC m=+145.476718530 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141003 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141025 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bc9m\" (UniqueName: \"kubernetes.io/projected/7898190a-5957-4772-abbb-1bfbf16f4eb5-kube-api-access-9bc9m\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141054 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz285\" (UniqueName: \"kubernetes.io/projected/806ad18b-e408-4679-9a37-b6efd48dd4d8-kube-api-access-bz285\") pod \"package-server-manager-789f6589d5-dp75h\" (UID: \"806ad18b-e408-4679-9a37-b6efd48dd4d8\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dp75h" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141074 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwxkv\" (UniqueName: \"kubernetes.io/projected/38703ec7-616b-499d-a173-4ea254e11d9a-kube-api-access-qwxkv\") pod \"dns-default-m9fbd\" (UID: \"38703ec7-616b-499d-a173-4ea254e11d9a\") " pod="openshift-dns/dns-default-m9fbd" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141111 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7898190a-5957-4772-abbb-1bfbf16f4eb5-audit-dir\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141125 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msqjg\" (UniqueName: \"kubernetes.io/projected/1a8d709a-565d-4213-ae67-17440c2d7678-kube-api-access-msqjg\") pod \"multus-admission-controller-857f4d67dd-zx8fv\" (UID: \"1a8d709a-565d-4213-ae67-17440c2d7678\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zx8fv" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141148 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-registry-tls\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141163 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d493be9f-6d12-4d4c-a04b-da608ff57eaa-serving-cert\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141179 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d493be9f-6d12-4d4c-a04b-da608ff57eaa-etcd-client\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141212 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d0d1509c-c26c-4066-8dab-23d0205d4211-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2qmsq\" (UID: \"d0d1509c-c26c-4066-8dab-23d0205d4211\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141257 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141273 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d493be9f-6d12-4d4c-a04b-da608ff57eaa-config\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141289 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnvvv\" (UniqueName: \"kubernetes.io/projected/1ae74e67-c9db-4ba6-87e5-0fcb923244b3-kube-api-access-cnvvv\") pod \"machine-config-controller-84d6567774-sx2jx\" (UID: \"1ae74e67-c9db-4ba6-87e5-0fcb923244b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141325 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6b88\" (UniqueName: \"kubernetes.io/projected/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-kube-api-access-g6b88\") pod \"collect-profiles-29417145-v956j\" (UID: \"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141351 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1ae74e67-c9db-4ba6-87e5-0fcb923244b3-proxy-tls\") pod \"machine-config-controller-84d6567774-sx2jx\" (UID: \"1ae74e67-c9db-4ba6-87e5-0fcb923244b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141370 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ed999748-8e6d-4402-81e3-1f483b8440e8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141387 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141414 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/532ce376-db1f-4770-a4cd-a30318242daf-console-serving-cert\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141430 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d0d1509c-c26c-4066-8dab-23d0205d4211-images\") pod \"machine-config-operator-74547568cd-2qmsq\" (UID: \"d0d1509c-c26c-4066-8dab-23d0205d4211\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141446 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzxhl\" (UniqueName: \"kubernetes.io/projected/aca05689-5cad-4e13-b513-1ef9de804ba9-kube-api-access-mzxhl\") pod \"catalog-operator-68c6474976-mfhhs\" (UID: \"aca05689-5cad-4e13-b513-1ef9de804ba9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141482 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed999748-8e6d-4402-81e3-1f483b8440e8-trusted-ca\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141498 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d493be9f-6d12-4d4c-a04b-da608ff57eaa-etcd-service-ca\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141527 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0b36426-e051-4864-b98e-b1ecf89914a4-config\") pod \"console-operator-58897d9998-t2255\" (UID: \"a0b36426-e051-4864-b98e-b1ecf89914a4\") " pod="openshift-console-operator/console-operator-58897d9998-t2255" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141542 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65a55c43-6a12-46dc-b181-8241ceab5d12-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-97565\" (UID: \"65a55c43-6a12-46dc-b181-8241ceab5d12\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-97565" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141558 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-secret-volume\") pod \"collect-profiles-29417145-v956j\" (UID: \"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141576 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-bound-sa-token\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141604 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/532ce376-db1f-4770-a4cd-a30318242daf-trusted-ca-bundle\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141637 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1a8d709a-565d-4213-ae67-17440c2d7678-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-zx8fv\" (UID: \"1a8d709a-565d-4213-ae67-17440c2d7678\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zx8fv" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141655 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141699 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/532ce376-db1f-4770-a4cd-a30318242daf-oauth-serving-cert\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141713 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0b36426-e051-4864-b98e-b1ecf89914a4-serving-cert\") pod \"console-operator-58897d9998-t2255\" (UID: \"a0b36426-e051-4864-b98e-b1ecf89914a4\") " pod="openshift-console-operator/console-operator-58897d9998-t2255" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141728 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-bkh56\" (UID: \"1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141744 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/532ce376-db1f-4770-a4cd-a30318242daf-console-config\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141768 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlrfv\" (UniqueName: \"kubernetes.io/projected/09f7b3f2-ef36-4a21-970a-3a19b7a6a726-kube-api-access-dlrfv\") pod \"packageserver-d55dfcdfc-d5zwg\" (UID: \"09f7b3f2-ef36-4a21-970a-3a19b7a6a726\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141795 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1ae74e67-c9db-4ba6-87e5-0fcb923244b3-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sx2jx\" (UID: \"1ae74e67-c9db-4ba6-87e5-0fcb923244b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141823 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141848 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d0d1509c-c26c-4066-8dab-23d0205d4211-proxy-tls\") pod \"machine-config-operator-74547568cd-2qmsq\" (UID: \"d0d1509c-c26c-4066-8dab-23d0205d4211\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141864 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-audit-policies\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141891 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghxlc\" (UniqueName: \"kubernetes.io/projected/1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90-kube-api-access-ghxlc\") pod \"cluster-image-registry-operator-dc59b4c8b-bkh56\" (UID: \"1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141913 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/65a55c43-6a12-46dc-b181-8241ceab5d12-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-97565\" (UID: \"65a55c43-6a12-46dc-b181-8241ceab5d12\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-97565" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141938 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141953 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141971 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.141986 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwm75\" (UniqueName: \"kubernetes.io/projected/532ce376-db1f-4770-a4cd-a30318242daf-kube-api-access-zwm75\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.142023 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ed999748-8e6d-4402-81e3-1f483b8440e8-registry-certificates\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.142039 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.142068 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/09f7b3f2-ef36-4a21-970a-3a19b7a6a726-webhook-cert\") pod \"packageserver-d55dfcdfc-d5zwg\" (UID: \"09f7b3f2-ef36-4a21-970a-3a19b7a6a726\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.142086 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/407db426-50cd-413d-bd5a-392aed7b1897-cert\") pod \"ingress-canary-kdfk5\" (UID: \"407db426-50cd-413d-bd5a-392aed7b1897\") " pod="openshift-ingress-canary/ingress-canary-kdfk5" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.142105 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxfnb\" (UniqueName: \"kubernetes.io/projected/d0d1509c-c26c-4066-8dab-23d0205d4211-kube-api-access-bxfnb\") pod \"machine-config-operator-74547568cd-2qmsq\" (UID: \"d0d1509c-c26c-4066-8dab-23d0205d4211\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.142123 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.142157 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/09f7b3f2-ef36-4a21-970a-3a19b7a6a726-apiservice-cert\") pod \"packageserver-d55dfcdfc-d5zwg\" (UID: \"09f7b3f2-ef36-4a21-970a-3a19b7a6a726\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.142185 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2daa96a6-fc64-4fd3-ba1a-d4ae6bbf97b0-metrics-tls\") pod \"dns-operator-744455d44c-cl6js\" (UID: \"2daa96a6-fc64-4fd3-ba1a-d4ae6bbf97b0\") " pod="openshift-dns-operator/dns-operator-744455d44c-cl6js" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.142200 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/09f7b3f2-ef36-4a21-970a-3a19b7a6a726-tmpfs\") pod \"packageserver-d55dfcdfc-d5zwg\" (UID: \"09f7b3f2-ef36-4a21-970a-3a19b7a6a726\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.142225 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trqvg\" (UniqueName: \"kubernetes.io/projected/2daa96a6-fc64-4fd3-ba1a-d4ae6bbf97b0-kube-api-access-trqvg\") pod \"dns-operator-744455d44c-cl6js\" (UID: \"2daa96a6-fc64-4fd3-ba1a-d4ae6bbf97b0\") " pod="openshift-dns-operator/dns-operator-744455d44c-cl6js" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.142259 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d493be9f-6d12-4d4c-a04b-da608ff57eaa-etcd-ca\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.142304 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a0b36426-e051-4864-b98e-b1ecf89914a4-trusted-ca\") pod \"console-operator-58897d9998-t2255\" (UID: \"a0b36426-e051-4864-b98e-b1ecf89914a4\") " pod="openshift-console-operator/console-operator-58897d9998-t2255" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.142329 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/38703ec7-616b-499d-a173-4ea254e11d9a-metrics-tls\") pod \"dns-default-m9fbd\" (UID: \"38703ec7-616b-499d-a173-4ea254e11d9a\") " pod="openshift-dns/dns-default-m9fbd" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.144295 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.145489 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ed999748-8e6d-4402-81e3-1f483b8440e8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.146014 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/aca05689-5cad-4e13-b513-1ef9de804ba9-profile-collector-cert\") pod \"catalog-operator-68c6474976-mfhhs\" (UID: \"aca05689-5cad-4e13-b513-1ef9de804ba9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.154876 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/09f7b3f2-ef36-4a21-970a-3a19b7a6a726-tmpfs\") pod \"packageserver-d55dfcdfc-d5zwg\" (UID: \"09f7b3f2-ef36-4a21-970a-3a19b7a6a726\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.154955 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.156033 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1a8d709a-565d-4213-ae67-17440c2d7678-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-zx8fv\" (UID: \"1a8d709a-565d-4213-ae67-17440c2d7678\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zx8fv" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.156180 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-config-volume\") pod \"collect-profiles-29417145-v956j\" (UID: \"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.156541 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d493be9f-6d12-4d4c-a04b-da608ff57eaa-etcd-ca\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.158128 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/532ce376-db1f-4770-a4cd-a30318242daf-oauth-serving-cert\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.158242 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d493be9f-6d12-4d4c-a04b-da608ff57eaa-etcd-client\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.158658 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7898190a-5957-4772-abbb-1bfbf16f4eb5-audit-dir\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.161679 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65a55c43-6a12-46dc-b181-8241ceab5d12-config\") pod \"kube-apiserver-operator-766d6c64bb-97565\" (UID: \"65a55c43-6a12-46dc-b181-8241ceab5d12\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-97565" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.167118 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/532ce376-db1f-4770-a4cd-a30318242daf-console-oauth-config\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.167488 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-audit-policies\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.168550 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.169104 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d493be9f-6d12-4d4c-a04b-da608ff57eaa-config\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.169481 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/532ce376-db1f-4770-a4cd-a30318242daf-console-config\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.171212 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ed999748-8e6d-4402-81e3-1f483b8440e8-registry-certificates\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.172294 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-bkh56\" (UID: \"1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.173399 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1ae74e67-c9db-4ba6-87e5-0fcb923244b3-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sx2jx\" (UID: \"1ae74e67-c9db-4ba6-87e5-0fcb923244b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.174050 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d0d1509c-c26c-4066-8dab-23d0205d4211-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2qmsq\" (UID: \"d0d1509c-c26c-4066-8dab-23d0205d4211\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.174733 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.181011 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.181071 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-secret-volume\") pod \"collect-profiles-29417145-v956j\" (UID: \"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.181874 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-bkh56\" (UID: \"1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.182546 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-registry-tls\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.182841 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/09f7b3f2-ef36-4a21-970a-3a19b7a6a726-webhook-cert\") pod \"packageserver-d55dfcdfc-d5zwg\" (UID: \"09f7b3f2-ef36-4a21-970a-3a19b7a6a726\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.182917 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.182973 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed999748-8e6d-4402-81e3-1f483b8440e8-trusted-ca\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.183232 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.183476 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d493be9f-6d12-4d4c-a04b-da608ff57eaa-etcd-service-ca\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.184165 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0b36426-e051-4864-b98e-b1ecf89914a4-config\") pod \"console-operator-58897d9998-t2255\" (UID: \"a0b36426-e051-4864-b98e-b1ecf89914a4\") " pod="openshift-console-operator/console-operator-58897d9998-t2255" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.184280 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/806ad18b-e408-4679-9a37-b6efd48dd4d8-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-dp75h\" (UID: \"806ad18b-e408-4679-9a37-b6efd48dd4d8\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dp75h" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.185141 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/532ce376-db1f-4770-a4cd-a30318242daf-service-ca\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.185331 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/532ce376-db1f-4770-a4cd-a30318242daf-trusted-ca-bundle\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.185537 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.186125 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d0d1509c-c26c-4066-8dab-23d0205d4211-images\") pod \"machine-config-operator-74547568cd-2qmsq\" (UID: \"d0d1509c-c26c-4066-8dab-23d0205d4211\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.186496 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d493be9f-6d12-4d4c-a04b-da608ff57eaa-serving-cert\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.188744 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2daa96a6-fc64-4fd3-ba1a-d4ae6bbf97b0-metrics-tls\") pod \"dns-operator-744455d44c-cl6js\" (UID: \"2daa96a6-fc64-4fd3-ba1a-d4ae6bbf97b0\") " pod="openshift-dns-operator/dns-operator-744455d44c-cl6js" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.189506 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65a55c43-6a12-46dc-b181-8241ceab5d12-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-97565\" (UID: \"65a55c43-6a12-46dc-b181-8241ceab5d12\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-97565" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.189567 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j86cv\" (UniqueName: \"kubernetes.io/projected/a0b36426-e051-4864-b98e-b1ecf89914a4-kube-api-access-j86cv\") pod \"console-operator-58897d9998-t2255\" (UID: \"a0b36426-e051-4864-b98e-b1ecf89914a4\") " pod="openshift-console-operator/console-operator-58897d9998-t2255" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.189897 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ed999748-8e6d-4402-81e3-1f483b8440e8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.190288 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d0d1509c-c26c-4066-8dab-23d0205d4211-proxy-tls\") pod \"machine-config-operator-74547568cd-2qmsq\" (UID: \"d0d1509c-c26c-4066-8dab-23d0205d4211\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.190408 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/09f7b3f2-ef36-4a21-970a-3a19b7a6a726-apiservice-cert\") pod \"packageserver-d55dfcdfc-d5zwg\" (UID: \"09f7b3f2-ef36-4a21-970a-3a19b7a6a726\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.191945 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0b36426-e051-4864-b98e-b1ecf89914a4-serving-cert\") pod \"console-operator-58897d9998-t2255\" (UID: \"a0b36426-e051-4864-b98e-b1ecf89914a4\") " pod="openshift-console-operator/console-operator-58897d9998-t2255" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.192799 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.193217 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/aca05689-5cad-4e13-b513-1ef9de804ba9-srv-cert\") pod \"catalog-operator-68c6474976-mfhhs\" (UID: \"aca05689-5cad-4e13-b513-1ef9de804ba9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.193347 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1ae74e67-c9db-4ba6-87e5-0fcb923244b3-proxy-tls\") pod \"machine-config-controller-84d6567774-sx2jx\" (UID: \"1ae74e67-c9db-4ba6-87e5-0fcb923244b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.194728 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.195037 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.201414 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/532ce376-db1f-4770-a4cd-a30318242daf-console-serving-cert\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.221756 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.230760 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-bkh56\" (UID: \"1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.233909 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwm75\" (UniqueName: \"kubernetes.io/projected/532ce376-db1f-4770-a4cd-a30318242daf-kube-api-access-zwm75\") pod \"console-f9d7485db-ldd62\" (UID: \"532ce376-db1f-4770-a4cd-a30318242daf\") " pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.233918 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-p8xpt"] Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.244294 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwxkv\" (UniqueName: \"kubernetes.io/projected/38703ec7-616b-499d-a173-4ea254e11d9a-kube-api-access-qwxkv\") pod \"dns-default-m9fbd\" (UID: \"38703ec7-616b-499d-a173-4ea254e11d9a\") " pod="openshift-dns/dns-default-m9fbd" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.245028 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/407db426-50cd-413d-bd5a-392aed7b1897-cert\") pod \"ingress-canary-kdfk5\" (UID: \"407db426-50cd-413d-bd5a-392aed7b1897\") " pod="openshift-ingress-canary/ingress-canary-kdfk5" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.245098 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/38703ec7-616b-499d-a173-4ea254e11d9a-metrics-tls\") pod \"dns-default-m9fbd\" (UID: \"38703ec7-616b-499d-a173-4ea254e11d9a\") " pod="openshift-dns/dns-default-m9fbd" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.245129 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.245176 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38703ec7-616b-499d-a173-4ea254e11d9a-config-volume\") pod \"dns-default-m9fbd\" (UID: \"38703ec7-616b-499d-a173-4ea254e11d9a\") " pod="openshift-dns/dns-default-m9fbd" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.245212 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctvsb\" (UniqueName: \"kubernetes.io/projected/407db426-50cd-413d-bd5a-392aed7b1897-kube-api-access-ctvsb\") pod \"ingress-canary-kdfk5\" (UID: \"407db426-50cd-413d-bd5a-392aed7b1897\") " pod="openshift-ingress-canary/ingress-canary-kdfk5" Dec 06 13:58:22 crc kubenswrapper[4706]: E1206 13:58:22.246803 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:22.746785662 +0000 UTC m=+145.582578519 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.248764 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38703ec7-616b-499d-a173-4ea254e11d9a-config-volume\") pod \"dns-default-m9fbd\" (UID: \"38703ec7-616b-499d-a173-4ea254e11d9a\") " pod="openshift-dns/dns-default-m9fbd" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.254425 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.256026 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trqvg\" (UniqueName: \"kubernetes.io/projected/2daa96a6-fc64-4fd3-ba1a-d4ae6bbf97b0-kube-api-access-trqvg\") pod \"dns-operator-744455d44c-cl6js\" (UID: \"2daa96a6-fc64-4fd3-ba1a-d4ae6bbf97b0\") " pod="openshift-dns-operator/dns-operator-744455d44c-cl6js" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.257787 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/38703ec7-616b-499d-a173-4ea254e11d9a-metrics-tls\") pod \"dns-default-m9fbd\" (UID: \"38703ec7-616b-499d-a173-4ea254e11d9a\") " pod="openshift-dns/dns-default-m9fbd" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.270595 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp5w4\" (UniqueName: \"kubernetes.io/projected/d493be9f-6d12-4d4c-a04b-da608ff57eaa-kube-api-access-dp5w4\") pod \"etcd-operator-b45778765-rlczt\" (UID: \"d493be9f-6d12-4d4c-a04b-da608ff57eaa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: W1206 13:58:22.280710 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0acf2c47_e9d7_4759_aa80_9f4ca8488899.slice/crio-490ce052580ef2b589ca5240c7ae89735489727d602a9ff466696090775216cf WatchSource:0}: Error finding container 490ce052580ef2b589ca5240c7ae89735489727d602a9ff466696090775216cf: Status 404 returned error can't find the container with id 490ce052580ef2b589ca5240c7ae89735489727d602a9ff466696090775216cf Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.283347 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2nk8h"] Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.286946 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msqjg\" (UniqueName: \"kubernetes.io/projected/1a8d709a-565d-4213-ae67-17440c2d7678-kube-api-access-msqjg\") pod \"multus-admission-controller-857f4d67dd-zx8fv\" (UID: \"1a8d709a-565d-4213-ae67-17440c2d7678\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zx8fv" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.291903 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg"] Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.292090 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/407db426-50cd-413d-bd5a-392aed7b1897-cert\") pod \"ingress-canary-kdfk5\" (UID: \"407db426-50cd-413d-bd5a-392aed7b1897\") " pod="openshift-ingress-canary/ingress-canary-kdfk5" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.297679 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-t2255" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.317157 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-cl6js" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.334726 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l25h5\" (UniqueName: \"kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-kube-api-access-l25h5\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.336011 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-zx8fv" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.340697 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlrfv\" (UniqueName: \"kubernetes.io/projected/09f7b3f2-ef36-4a21-970a-3a19b7a6a726-kube-api-access-dlrfv\") pod \"packageserver-d55dfcdfc-d5zwg\" (UID: \"09f7b3f2-ef36-4a21-970a-3a19b7a6a726\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.348010 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:22 crc kubenswrapper[4706]: E1206 13:58:22.348475 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:22.848455504 +0000 UTC m=+145.684248361 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.358717 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghxlc\" (UniqueName: \"kubernetes.io/projected/1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90-kube-api-access-ghxlc\") pod \"cluster-image-registry-operator-dc59b4c8b-bkh56\" (UID: \"1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.373640 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.404730 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.420285 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnvvv\" (UniqueName: \"kubernetes.io/projected/1ae74e67-c9db-4ba6-87e5-0fcb923244b3-kube-api-access-cnvvv\") pod \"machine-config-controller-84d6567774-sx2jx\" (UID: \"1ae74e67-c9db-4ba6-87e5-0fcb923244b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.422352 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/65a55c43-6a12-46dc-b181-8241ceab5d12-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-97565\" (UID: \"65a55c43-6a12-46dc-b181-8241ceab5d12\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-97565" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.424376 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" event={"ID":"0136a364-d7ad-445a-9226-5b064e4b4527","Type":"ContainerStarted","Data":"a2901dd6a8fc0b8093df7d8b225c18dc7b49d40ea2b81f470c69da25b7db1412"} Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.426318 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" event={"ID":"0acf2c47-e9d7-4759-aa80-9f4ca8488899","Type":"ContainerStarted","Data":"490ce052580ef2b589ca5240c7ae89735489727d602a9ff466696090775216cf"} Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.432703 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9jn7l" event={"ID":"621a48da-c965-4d74-abd0-62e119fd81c8","Type":"ContainerStarted","Data":"4854de3b6b3c33347e59e53e60c3bc08c969849672b264784e2b883483b75d46"} Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.441511 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6b88\" (UniqueName: \"kubernetes.io/projected/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-kube-api-access-g6b88\") pod \"collect-profiles-29417145-v956j\" (UID: \"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.445838 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r94jx" event={"ID":"b9f16df2-5700-43c2-9488-78c3a9d8bcd6","Type":"ContainerStarted","Data":"3097d60f3d6fcb9ec8c512630542cafa31e1a9eb50f49c000e6ddff50b3b7075"} Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.449178 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p85f8" event={"ID":"77b4e41c-830d-48c2-8d00-97b33c7e8a29","Type":"ContainerStarted","Data":"316c6234af61df4305df9aa3438f10bfa5e03b6d9febe56ce1265c4c33fc9489"} Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.458969 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: E1206 13:58:22.459326 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:22.959299372 +0000 UTC m=+145.795092229 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.463162 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bc9m\" (UniqueName: \"kubernetes.io/projected/7898190a-5957-4772-abbb-1bfbf16f4eb5-kube-api-access-9bc9m\") pod \"oauth-openshift-558db77b4-l55q6\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.467212 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz285\" (UniqueName: \"kubernetes.io/projected/806ad18b-e408-4679-9a37-b6efd48dd4d8-kube-api-access-bz285\") pod \"package-server-manager-789f6589d5-dp75h\" (UID: \"806ad18b-e408-4679-9a37-b6efd48dd4d8\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dp75h" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.471871 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" event={"ID":"5cc40c37-833b-439a-bfd3-b6aa63b8c033","Type":"ContainerStarted","Data":"0aa07d2afb61438e1394d713fb2cdd4ac3294f665693ad420be21772c42802a5"} Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.511785 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-bound-sa-token\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.511786 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxfnb\" (UniqueName: \"kubernetes.io/projected/d0d1509c-c26c-4066-8dab-23d0205d4211-kube-api-access-bxfnb\") pod \"machine-config-operator-74547568cd-2qmsq\" (UID: \"d0d1509c-c26c-4066-8dab-23d0205d4211\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.528442 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzxhl\" (UniqueName: \"kubernetes.io/projected/aca05689-5cad-4e13-b513-1ef9de804ba9-kube-api-access-mzxhl\") pod \"catalog-operator-68c6474976-mfhhs\" (UID: \"aca05689-5cad-4e13-b513-1ef9de804ba9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.530080 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-j4chj"] Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.533562 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctvsb\" (UniqueName: \"kubernetes.io/projected/407db426-50cd-413d-bd5a-392aed7b1897-kube-api-access-ctvsb\") pod \"ingress-canary-kdfk5\" (UID: \"407db426-50cd-413d-bd5a-392aed7b1897\") " pod="openshift-ingress-canary/ingress-canary-kdfk5" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.536790 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.537317 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cvtpw"] Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.558689 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwxkv\" (UniqueName: \"kubernetes.io/projected/38703ec7-616b-499d-a173-4ea254e11d9a-kube-api-access-qwxkv\") pod \"dns-default-m9fbd\" (UID: \"38703ec7-616b-499d-a173-4ea254e11d9a\") " pod="openshift-dns/dns-default-m9fbd" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.569240 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.569767 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:22 crc kubenswrapper[4706]: E1206 13:58:22.569884 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:23.069856302 +0000 UTC m=+145.905649159 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.570154 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: E1206 13:58:22.572407 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:23.072395848 +0000 UTC m=+145.908188695 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.607598 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.618686 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.628078 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.649604 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.650649 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-97565" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.654909 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dp75h" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.685099 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" podStartSLOduration=126.685085675 podStartE2EDuration="2m6.685085675s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:22.684503669 +0000 UTC m=+145.520296526" watchObservedRunningTime="2025-12-06 13:58:22.685085675 +0000 UTC m=+145.520878532" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.686925 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:22 crc kubenswrapper[4706]: E1206 13:58:22.687275 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:23.187260371 +0000 UTC m=+146.023053228 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.688398 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-j2hql"] Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.696864 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.765702 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kdfk5" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.769364 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-m9fbd" Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.788426 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: E1206 13:58:22.789184 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:23.289169058 +0000 UTC m=+146.124961915 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.819831 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-595n2"] Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.889458 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:22 crc kubenswrapper[4706]: E1206 13:58:22.889772 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:23.389752461 +0000 UTC m=+146.225545308 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.889819 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:22 crc kubenswrapper[4706]: E1206 13:58:22.890189 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:23.390168572 +0000 UTC m=+146.225961429 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:22 crc kubenswrapper[4706]: I1206 13:58:22.996189 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:22 crc kubenswrapper[4706]: E1206 13:58:22.996564 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:23.496549045 +0000 UTC m=+146.332341902 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.000528 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-9hxp8" podStartSLOduration=127.000505128 podStartE2EDuration="2m7.000505128s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:22.995755325 +0000 UTC m=+145.831548172" watchObservedRunningTime="2025-12-06 13:58:23.000505128 +0000 UTC m=+145.836297995" Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.051646 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp64k"] Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.075450 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8"] Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.108481 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:23 crc kubenswrapper[4706]: E1206 13:58:23.109277 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:23.609259782 +0000 UTC m=+146.445052639 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.211655 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:23 crc kubenswrapper[4706]: E1206 13:58:23.212124 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:23.712105674 +0000 UTC m=+146.547898531 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.257787 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qgdc9" podStartSLOduration=127.257768096 podStartE2EDuration="2m7.257768096s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:23.255373864 +0000 UTC m=+146.091166721" watchObservedRunningTime="2025-12-06 13:58:23.257768096 +0000 UTC m=+146.093560953" Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.313545 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:23 crc kubenswrapper[4706]: E1206 13:58:23.314148 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:23.814133474 +0000 UTC m=+146.649926331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.352772 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr"] Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.364860 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8h4wk"] Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.415589 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:23 crc kubenswrapper[4706]: E1206 13:58:23.415986 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:23.91597055 +0000 UTC m=+146.751763407 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:23 crc kubenswrapper[4706]: W1206 13:58:23.464936 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6f43404_a6e1_49ee_bbef_9e027779cef5.slice/crio-af2f376c2f3b50cf439692c0e4f5103103fef580d757a5917b32b658b8533a1c WatchSource:0}: Error finding container af2f376c2f3b50cf439692c0e4f5103103fef580d757a5917b32b658b8533a1c: Status 404 returned error can't find the container with id af2f376c2f3b50cf439692c0e4f5103103fef580d757a5917b32b658b8533a1c Dec 06 13:58:23 crc kubenswrapper[4706]: W1206 13:58:23.465302 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8de762a2_b5b9_4e05_aedc_9ba4d05ee95f.slice/crio-9fbe8dd068be63657f5d8bf324425291a8bc955138b5a5fdf2850bc4e28c4920 WatchSource:0}: Error finding container 9fbe8dd068be63657f5d8bf324425291a8bc955138b5a5fdf2850bc4e28c4920: Status 404 returned error can't find the container with id 9fbe8dd068be63657f5d8bf324425291a8bc955138b5a5fdf2850bc4e28c4920 Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.498664 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-r2t78"] Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.502183 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-k9rh4"] Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.529483 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:23 crc kubenswrapper[4706]: E1206 13:58:23.529829 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.029818586 +0000 UTC m=+146.865611433 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.554320 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r94jx" event={"ID":"b9f16df2-5700-43c2-9488-78c3a9d8bcd6","Type":"ContainerStarted","Data":"6ead0c26055d1e97a8aee34fd6d5af40d00af676ad2f2ab43ec35bc6b4e15d0c"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.554365 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-595n2" event={"ID":"b38e2a11-98f9-47c3-a711-106d6aa308de","Type":"ContainerStarted","Data":"bc76f69ffdd4775c7195a6c2593e1cffd07cb03397ca6422832b8773c2768c4f"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.584087 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" event={"ID":"fe23e3f5-57af-4fd2-a8bd-9364bbe39776","Type":"ContainerStarted","Data":"ed6d1c472bd8573f4497ae9a81eac0289f4e1d407bf0f2a5c59fd7571b774ef0"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.601645 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2"] Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.630045 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:23 crc kubenswrapper[4706]: E1206 13:58:23.631458 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.131442766 +0000 UTC m=+146.967235623 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.697246 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" event={"ID":"5cc40c37-833b-439a-bfd3-b6aa63b8c033","Type":"ContainerStarted","Data":"f2b68c0870d32771335b81d1782a07ff8cccc1246c2d343930b3c033d05cdca8"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.703145 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ldd62"] Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.727711 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4"] Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.731462 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:23 crc kubenswrapper[4706]: E1206 13:58:23.731798 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.231786583 +0000 UTC m=+147.067579440 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.775766 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" event={"ID":"caf8bd48-353a-428e-ba7a-1a7c40f19b23","Type":"ContainerStarted","Data":"a07e12a234438187fce32d8419783fe9a44d6cea95d75bf2676f88b2891bae9f"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.775963 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" event={"ID":"caf8bd48-353a-428e-ba7a-1a7c40f19b23","Type":"ContainerStarted","Data":"43abec08a7c11fd95774a38e1e7e4b0a61e23913b70ec61ae88aa4e59d75dc22"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.802107 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-zx8fv"] Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.844847 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:23 crc kubenswrapper[4706]: E1206 13:58:23.844972 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.344956002 +0000 UTC m=+147.180748859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.845554 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:23 crc kubenswrapper[4706]: E1206 13:58:23.846023 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.345968838 +0000 UTC m=+147.181761695 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.846555 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-nngbv" event={"ID":"6b5dec00-19df-4fb7-883e-6941ebe98d0a","Type":"ContainerStarted","Data":"019216153d5f0fbc17ee150c2cb31f4aad2f8132673351a0aad435998f4566a6"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.846603 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-nngbv" event={"ID":"6b5dec00-19df-4fb7-883e-6941ebe98d0a","Type":"ContainerStarted","Data":"85159f481a7703589e046327be4f62aff03c0a911f547dfacb227532c1bebf17"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.852178 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cl6js"] Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.860899 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" event={"ID":"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b","Type":"ContainerStarted","Data":"1e8fda098971671008f1b8749d30e747e0444484738057bda313edb0af94698a"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.863247 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9jn7l" event={"ID":"621a48da-c965-4d74-abd0-62e119fd81c8","Type":"ContainerStarted","Data":"9a802635134bfe63b76351fe292683fde8e4d777308cd36b72cb740c968ea9c0"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.863620 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-9jn7l" Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.864208 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rlczt"] Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.864940 4706 patch_prober.go:28] interesting pod/downloads-7954f5f757-9jn7l container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.865000 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9jn7l" podUID="621a48da-c965-4d74-abd0-62e119fd81c8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.869757 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-t2255"] Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.872533 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2nk8h" event={"ID":"bc8af660-99b4-4d61-bc9c-4580f285d6c6","Type":"ContainerStarted","Data":"1c92f79a09aa2c85f2bdff2fdcc41bfdfc2016743e374969c81426c31fcf82da"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.872572 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2nk8h" event={"ID":"bc8af660-99b4-4d61-bc9c-4580f285d6c6","Type":"ContainerStarted","Data":"8e1b3b7142316da937d51a38c16acfe4c05f6895655bff1b2bbd06fdd9a21d16"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.880839 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg"] Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.882370 4706 generic.go:334] "Generic (PLEG): container finished" podID="0acf2c47-e9d7-4759-aa80-9f4ca8488899" containerID="5e91ecc6b2ebe97d04251ac048acad5ef28017db4418d904bcfc4fb4caed0481" exitCode=0 Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.882420 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" event={"ID":"0acf2c47-e9d7-4759-aa80-9f4ca8488899","Type":"ContainerDied","Data":"5e91ecc6b2ebe97d04251ac048acad5ef28017db4418d904bcfc4fb4caed0481"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.909900 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p85f8" event={"ID":"77b4e41c-830d-48c2-8d00-97b33c7e8a29","Type":"ContainerStarted","Data":"6a7c158e7a5710fb388933ebd802590fc66f43a5efd257fdbd6fb0e2f4e9603b"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.921901 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" event={"ID":"f6f43404-a6e1-49ee-bbef-9e027779cef5","Type":"ContainerStarted","Data":"af2f376c2f3b50cf439692c0e4f5103103fef580d757a5917b32b658b8533a1c"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.923831 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp64k" event={"ID":"2fab8c56-bd27-485d-8216-c219effe9dcc","Type":"ContainerStarted","Data":"38d1ae98db89c735a8f99cfcd0f8282ca3fd81fc78e33d1052900f06cdcb461e"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.925481 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" event={"ID":"b801f158-2df3-44c7-8bfe-5caa8467e29d","Type":"ContainerStarted","Data":"a4c30054331e61c6b1cb168dbe02d6eedccf1da77ed92b65306384af25abbddd"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.927503 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-j2hql" event={"ID":"f06c4ffb-6d3f-4189-8e4e-4d5233c41faa","Type":"ContainerStarted","Data":"d0ea284e94cc89221a0f4b3140736ff77c0e05711e272c4d207dd4d9e6da649c"} Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.930682 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.946471 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:23 crc kubenswrapper[4706]: E1206 13:58:23.947510 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.447492706 +0000 UTC m=+147.283285563 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:23 crc kubenswrapper[4706]: I1206 13:58:23.949668 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8h4wk" event={"ID":"8de762a2-b5b9-4e05-aedc-9ba4d05ee95f","Type":"ContainerStarted","Data":"9fbe8dd068be63657f5d8bf324425291a8bc955138b5a5fdf2850bc4e28c4920"} Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.026973 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" podStartSLOduration=127.026953812 podStartE2EDuration="2m7.026953812s" podCreationTimestamp="2025-12-06 13:56:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:24.008112044 +0000 UTC m=+146.843904911" watchObservedRunningTime="2025-12-06 13:58:24.026953812 +0000 UTC m=+146.862746669" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.045515 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:24 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:24 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:24 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.045573 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.047432 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:24 crc kubenswrapper[4706]: E1206 13:58:24.051891 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.551879227 +0000 UTC m=+147.387672084 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:24 crc kubenswrapper[4706]: W1206 13:58:24.102423 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2daa96a6_fc64_4fd3_ba1a_d4ae6bbf97b0.slice/crio-946505ba43511a36ad5ec2f92313a154b5272751641cf3d317eac966cd7d4051 WatchSource:0}: Error finding container 946505ba43511a36ad5ec2f92313a154b5272751641cf3d317eac966cd7d4051: Status 404 returned error can't find the container with id 946505ba43511a36ad5ec2f92313a154b5272751641cf3d317eac966cd7d4051 Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.148850 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:24 crc kubenswrapper[4706]: E1206 13:58:24.149017 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.648998101 +0000 UTC m=+147.484790958 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.149047 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:24 crc kubenswrapper[4706]: E1206 13:58:24.149397 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.649389401 +0000 UTC m=+147.485182248 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.182741 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-j2hql" podStartSLOduration=127.182725534 podStartE2EDuration="2m7.182725534s" podCreationTimestamp="2025-12-06 13:56:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:24.182425836 +0000 UTC m=+147.018218693" watchObservedRunningTime="2025-12-06 13:58:24.182725534 +0000 UTC m=+147.018518381" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.183345 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-9jn7l" podStartSLOduration=128.183339589 podStartE2EDuration="2m8.183339589s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:24.129084635 +0000 UTC m=+146.964877482" watchObservedRunningTime="2025-12-06 13:58:24.183339589 +0000 UTC m=+147.019132446" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.213819 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56"] Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.218866 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dp75h"] Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.225765 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p85f8" podStartSLOduration=128.225750127 podStartE2EDuration="2m8.225750127s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:24.223408567 +0000 UTC m=+147.059201424" watchObservedRunningTime="2025-12-06 13:58:24.225750127 +0000 UTC m=+147.061542984" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.249992 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:24 crc kubenswrapper[4706]: E1206 13:58:24.250178 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.750150899 +0000 UTC m=+147.585943756 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.250217 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:24 crc kubenswrapper[4706]: E1206 13:58:24.250580 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.750567409 +0000 UTC m=+147.586360266 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.293692 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-nngbv" podStartSLOduration=5.293676985 podStartE2EDuration="5.293676985s" podCreationTimestamp="2025-12-06 13:58:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:24.289488837 +0000 UTC m=+147.125281684" watchObservedRunningTime="2025-12-06 13:58:24.293676985 +0000 UTC m=+147.129469842" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.355086 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:24 crc kubenswrapper[4706]: E1206 13:58:24.355414 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.855399242 +0000 UTC m=+147.691192099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.363148 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx"] Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.375210 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2nk8h" podStartSLOduration=128.375194555 podStartE2EDuration="2m8.375194555s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:24.374714262 +0000 UTC m=+147.210507109" watchObservedRunningTime="2025-12-06 13:58:24.375194555 +0000 UTC m=+147.210987412" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.439039 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-97565"] Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.462335 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.462386 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.462429 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.462455 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.462488 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.465121 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq"] Dec 06 13:58:24 crc kubenswrapper[4706]: E1206 13:58:24.468759 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:24.968743606 +0000 UTC m=+147.804536463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.470288 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j"] Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.484969 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.493974 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.498395 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.498434 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.505950 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-r94jx" podStartSLOduration=128.505935218 podStartE2EDuration="2m8.505935218s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:24.498244729 +0000 UTC m=+147.334037586" watchObservedRunningTime="2025-12-06 13:58:24.505935218 +0000 UTC m=+147.341728075" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.507910 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kdfk5"] Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.518048 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l55q6"] Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.526374 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7xn8m"] Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.563675 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:24 crc kubenswrapper[4706]: E1206 13:58:24.564477 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:25.064459803 +0000 UTC m=+147.900252660 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.596093 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-m9fbd"] Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.597892 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs"] Dec 06 13:58:24 crc kubenswrapper[4706]: W1206 13:58:24.617098 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65a55c43_6a12_46dc_b181_8241ceab5d12.slice/crio-7f55da1197978167eca10248c34d1f6f5df86d387c26be3bbf6c0ce19127522b WatchSource:0}: Error finding container 7f55da1197978167eca10248c34d1f6f5df86d387c26be3bbf6c0ce19127522b: Status 404 returned error can't find the container with id 7f55da1197978167eca10248c34d1f6f5df86d387c26be3bbf6c0ce19127522b Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.636814 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.646652 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 13:58:24 crc kubenswrapper[4706]: W1206 13:58:24.661695 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded3cbb3f_fda7_48cb_baa4_c6d1dc7257b3.slice/crio-23e9db62230960184342a1cd4ca8f54fcbde0688a07f952b12632534cdf01351 WatchSource:0}: Error finding container 23e9db62230960184342a1cd4ca8f54fcbde0688a07f952b12632534cdf01351: Status 404 returned error can't find the container with id 23e9db62230960184342a1cd4ca8f54fcbde0688a07f952b12632534cdf01351 Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.665292 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:24 crc kubenswrapper[4706]: E1206 13:58:24.665546 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:25.165536869 +0000 UTC m=+148.001329716 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:24 crc kubenswrapper[4706]: W1206 13:58:24.693483 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0d1509c_c26c_4066_8dab_23d0205d4211.slice/crio-6252e3b84717123dbd1f22b96242df64665cd7edf340a71c859864d91da8f4a6 WatchSource:0}: Error finding container 6252e3b84717123dbd1f22b96242df64665cd7edf340a71c859864d91da8f4a6: Status 404 returned error can't find the container with id 6252e3b84717123dbd1f22b96242df64665cd7edf340a71c859864d91da8f4a6 Dec 06 13:58:24 crc kubenswrapper[4706]: W1206 13:58:24.702580 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7898190a_5957_4772_abbb_1bfbf16f4eb5.slice/crio-d5109d0031068713991c9cf3045d11f4a0988baf1963dc4cd68791a06fe6d7ea WatchSource:0}: Error finding container d5109d0031068713991c9cf3045d11f4a0988baf1963dc4cd68791a06fe6d7ea: Status 404 returned error can't find the container with id d5109d0031068713991c9cf3045d11f4a0988baf1963dc4cd68791a06fe6d7ea Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.751711 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.772090 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:24 crc kubenswrapper[4706]: E1206 13:58:24.774495 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:25.274461758 +0000 UTC m=+148.110254615 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.876023 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:24 crc kubenswrapper[4706]: E1206 13:58:24.876379 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:25.376366985 +0000 UTC m=+148.212159842 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.936774 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:24 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:24 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:24 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.936807 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.973762 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" event={"ID":"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b","Type":"ContainerStarted","Data":"f7e97b997ea0bad25bb3f00476367a46da5d7d8a15abcdcdefaa0039a4cc2282"} Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.974194 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.978229 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:24 crc kubenswrapper[4706]: E1206 13:58:24.978685 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:25.478665193 +0000 UTC m=+148.314458050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.978832 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-595n2" event={"ID":"b38e2a11-98f9-47c3-a711-106d6aa308de","Type":"ContainerStarted","Data":"39cf078d718af1337268f8181d92c71396be986ede6c49d101474d819591641b"} Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.978863 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-595n2" event={"ID":"b38e2a11-98f9-47c3-a711-106d6aa308de","Type":"ContainerStarted","Data":"233671bc8299a0177d31de9bdc485b90973043d3d458989563a8805494512e7f"} Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.980993 4706 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cvtpw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.981063 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.985801 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" event={"ID":"d493be9f-6d12-4d4c-a04b-da608ff57eaa","Type":"ContainerStarted","Data":"631231f3d2809d157e3726b5d343f521a88d65ad5174ff4bdbbaa0787d110596"} Dec 06 13:58:24 crc kubenswrapper[4706]: I1206 13:58:24.985836 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" event={"ID":"d493be9f-6d12-4d4c-a04b-da608ff57eaa","Type":"ContainerStarted","Data":"f95e1828788536e5aa4aba519a78997feb35d1854dacb9522bb0dc1784903d60"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.001506 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-k9rh4" event={"ID":"56c2d16b-fba2-4c54-95fb-71e7676b4445","Type":"ContainerStarted","Data":"d936a0a4ef8ad00f3c8bf91071ba21c93f2767096adce10734c05463bdaa2443"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.001543 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-k9rh4" event={"ID":"56c2d16b-fba2-4c54-95fb-71e7676b4445","Type":"ContainerStarted","Data":"31940b8ee834086c7d5951e050b0af4ff0daaad48b1b0d0a404cbd1b8f184709"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.012710 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" event={"ID":"b801f158-2df3-44c7-8bfe-5caa8467e29d","Type":"ContainerStarted","Data":"41e759f55161abf22d950b30fa767aeed37879f3431c5aacf9f23cbdc1040c2a"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.014483 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" event={"ID":"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3","Type":"ContainerStarted","Data":"23e9db62230960184342a1cd4ca8f54fcbde0688a07f952b12632534cdf01351"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.017033 4706 generic.go:334] "Generic (PLEG): container finished" podID="caf8bd48-353a-428e-ba7a-1a7c40f19b23" containerID="a07e12a234438187fce32d8419783fe9a44d6cea95d75bf2676f88b2891bae9f" exitCode=0 Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.017085 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" event={"ID":"caf8bd48-353a-428e-ba7a-1a7c40f19b23","Type":"ContainerDied","Data":"a07e12a234438187fce32d8419783fe9a44d6cea95d75bf2676f88b2891bae9f"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.017105 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" event={"ID":"caf8bd48-353a-428e-ba7a-1a7c40f19b23","Type":"ContainerStarted","Data":"1bed5d2f6fde19729f1f7dbe1a9b119a5ad213eb59d64949c0e4ce48bad6c4be"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.017795 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.026052 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx" event={"ID":"1ae74e67-c9db-4ba6-87e5-0fcb923244b3","Type":"ContainerStarted","Data":"dda533aab5561e53e1cae994c90275f3f07d0973efa66c798adfa0207bfb270a"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.035276 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" podStartSLOduration=128.035259687 podStartE2EDuration="2m8.035259687s" podCreationTimestamp="2025-12-06 13:56:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:24.997729826 +0000 UTC m=+147.833522683" watchObservedRunningTime="2025-12-06 13:58:25.035259687 +0000 UTC m=+147.871052544" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.036476 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-595n2" podStartSLOduration=129.036472269 podStartE2EDuration="2m9.036472269s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:25.034804756 +0000 UTC m=+147.870597603" watchObservedRunningTime="2025-12-06 13:58:25.036472269 +0000 UTC m=+147.872265116" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.057573 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2" event={"ID":"9f1fce81-bf5b-48df-a92a-e95e82b5ef54","Type":"ContainerStarted","Data":"bfa503283c3d945de66ddd3a9a502ca0e1fcf238bc09a5fd29c0ce752cefcd00"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.057615 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2" event={"ID":"9f1fce81-bf5b-48df-a92a-e95e82b5ef54","Type":"ContainerStarted","Data":"d3bdc1a3738525350520bf94535edf99071aacb790974590a177d68fbbd5cd63"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.061243 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2t78" event={"ID":"d2b3bc37-36b2-473e-ae10-b0bdc30f2ef2","Type":"ContainerStarted","Data":"b8d3b37e02455eb9847c096f8f4db836aba7076b2baa45e76c86f3c7508bf48a"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.061265 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2t78" event={"ID":"d2b3bc37-36b2-473e-ae10-b0bdc30f2ef2","Type":"ContainerStarted","Data":"c448f5a9f82dcd6e2444b65de982dfb3dabb3ea4faf713c2987ed70a9f1d8519"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.065422 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-rlczt" podStartSLOduration=129.065403057 podStartE2EDuration="2m9.065403057s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:25.064551525 +0000 UTC m=+147.900344382" watchObservedRunningTime="2025-12-06 13:58:25.065403057 +0000 UTC m=+147.901195914" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.068359 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp64k" event={"ID":"2fab8c56-bd27-485d-8216-c219effe9dcc","Type":"ContainerStarted","Data":"da13915a00fa57b1b0159f5ef00eeb64f4f052a6fc217e6cbe2b0d705b06c117"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.079073 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" event={"ID":"27733066-fbcd-474e-9af2-d2bd168533c4","Type":"ContainerStarted","Data":"942d1a9d573510edd56532c1286830d479e71c044a6ef44b945772c8885d09f2"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.079972 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:25 crc kubenswrapper[4706]: E1206 13:58:25.080221 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:25.580211291 +0000 UTC m=+148.416004148 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.082989 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" event={"ID":"0acf2c47-e9d7-4759-aa80-9f4ca8488899","Type":"ContainerStarted","Data":"bf04a4697abf93e068cf6c75c541f7ac8fa9cb6b928e466803bb209b9b5457f7"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.093901 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-t2255" event={"ID":"a0b36426-e051-4864-b98e-b1ecf89914a4","Type":"ContainerStarted","Data":"b77afa0d2d1295fc19d06d3b4c5ec36f6d6d394484985c6ee7e8ac0f32e1f99e"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.094855 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-t2255" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.100096 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" podStartSLOduration=129.100077975 podStartE2EDuration="2m9.100077975s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:25.097215051 +0000 UTC m=+147.933007898" watchObservedRunningTime="2025-12-06 13:58:25.100077975 +0000 UTC m=+147.935870832" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.100242 4706 patch_prober.go:28] interesting pod/console-operator-58897d9998-t2255 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/readyz\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.100296 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-t2255" podUID="a0b36426-e051-4864-b98e-b1ecf89914a4" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.28:8443/readyz\": dial tcp 10.217.0.28:8443: connect: connection refused" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.113968 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-j2hql" event={"ID":"f06c4ffb-6d3f-4189-8e4e-4d5233c41faa","Type":"ContainerStarted","Data":"66c567c65cd190b8e1bb6f2c613def7c3f367c170746713f12d1689033ffc7c9"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.126675 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs" event={"ID":"aca05689-5cad-4e13-b513-1ef9de804ba9","Type":"ContainerStarted","Data":"6dd6303cbdec7087fc2e6749d197dda236315eafbe6636410020c850eccac216"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.147983 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kdfk5" event={"ID":"407db426-50cd-413d-bd5a-392aed7b1897","Type":"ContainerStarted","Data":"b6fc10af3fcffa5528d6b97eb414fe71888924ccaa6881d50a65d5b150f4eacf"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.165915 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" podStartSLOduration=128.165892088 podStartE2EDuration="2m8.165892088s" podCreationTimestamp="2025-12-06 13:56:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:25.13738833 +0000 UTC m=+147.973181187" watchObservedRunningTime="2025-12-06 13:58:25.165892088 +0000 UTC m=+148.001684945" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.166274 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-97565" event={"ID":"65a55c43-6a12-46dc-b181-8241ceab5d12","Type":"ContainerStarted","Data":"7f55da1197978167eca10248c34d1f6f5df86d387c26be3bbf6c0ce19127522b"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.180880 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:25 crc kubenswrapper[4706]: E1206 13:58:25.181887 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:25.681870382 +0000 UTC m=+148.517663239 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.183513 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8h4wk" event={"ID":"8de762a2-b5b9-4e05-aedc-9ba4d05ee95f","Type":"ContainerStarted","Data":"be6441df87df13c7ce63083d487574859930545dfb43274e25122d25f4d5b295"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.190051 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56" event={"ID":"1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90","Type":"ContainerStarted","Data":"88c0b5ff8f4860f73537e1b6bea1b30d6d0b0440a2070f0a6afa2d10e2dc1f1a"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.193991 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-k9rh4" podStartSLOduration=128.193979455 podStartE2EDuration="2m8.193979455s" podCreationTimestamp="2025-12-06 13:56:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:25.193264746 +0000 UTC m=+148.029057614" watchObservedRunningTime="2025-12-06 13:58:25.193979455 +0000 UTC m=+148.029772312" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.194640 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-zx8fv" event={"ID":"1a8d709a-565d-4213-ae67-17440c2d7678","Type":"ContainerStarted","Data":"f60f723ec3d685b6c945a650305c2c6eaf2d02ab8f71244307c6ee0801e680e1"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.194684 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-zx8fv" event={"ID":"1a8d709a-565d-4213-ae67-17440c2d7678","Type":"ContainerStarted","Data":"0a4ea3b8f89e7c14898a7bad8489d074420b23410b02a086e97e63518be9a208"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.195202 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6hsl2" podStartSLOduration=129.195194906 podStartE2EDuration="2m9.195194906s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:25.165633931 +0000 UTC m=+148.001426788" watchObservedRunningTime="2025-12-06 13:58:25.195194906 +0000 UTC m=+148.030987763" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.197153 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" event={"ID":"5cc40c37-833b-439a-bfd3-b6aa63b8c033","Type":"ContainerStarted","Data":"520a5b79799717495ae4109d80e78569ec0ade4979ff94fbbbd047feef2a7b37"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.204060 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" event={"ID":"d0d1509c-c26c-4066-8dab-23d0205d4211","Type":"ContainerStarted","Data":"6252e3b84717123dbd1f22b96242df64665cd7edf340a71c859864d91da8f4a6"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.208066 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" event={"ID":"7898190a-5957-4772-abbb-1bfbf16f4eb5","Type":"ContainerStarted","Data":"d5109d0031068713991c9cf3045d11f4a0988baf1963dc4cd68791a06fe6d7ea"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.220621 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4" event={"ID":"f65fc54c-f0ca-40dc-a8bc-9084483d2d11","Type":"ContainerStarted","Data":"c7631cdc38e02e3f9bc4b598de95a8c0cdd8522de96cc26f51307c035ad0d57e"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.221654 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.230403 4706 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-gfkh4 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.230454 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4" podUID="f65fc54c-f0ca-40dc-a8bc-9084483d2d11" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.263428 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8h4wk" podStartSLOduration=129.263411352 podStartE2EDuration="2m9.263411352s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:25.226069815 +0000 UTC m=+148.061862672" watchObservedRunningTime="2025-12-06 13:58:25.263411352 +0000 UTC m=+148.099204209" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.264371 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp64k" podStartSLOduration=129.264365217 podStartE2EDuration="2m9.264365217s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:25.263608887 +0000 UTC m=+148.099401744" watchObservedRunningTime="2025-12-06 13:58:25.264365217 +0000 UTC m=+148.100158074" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.272084 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" event={"ID":"09f7b3f2-ef36-4a21-970a-3a19b7a6a726","Type":"ContainerStarted","Data":"5dc1d87b1b2e31265e5d2518e5cd28af07876db45d2c3319214bc62b327bf1b4"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.272874 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.285022 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:25 crc kubenswrapper[4706]: E1206 13:58:25.286777 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:25.786764316 +0000 UTC m=+148.622557173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.287506 4706 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-d5zwg container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:5443/healthz\": dial tcp 10.217.0.31:5443: connect: connection refused" start-of-body= Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.287567 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" podUID="09f7b3f2-ef36-4a21-970a-3a19b7a6a726" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.31:5443/healthz\": dial tcp 10.217.0.31:5443: connect: connection refused" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.289354 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-t2255" podStartSLOduration=129.289341383 podStartE2EDuration="2m9.289341383s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:25.287987428 +0000 UTC m=+148.123780275" watchObservedRunningTime="2025-12-06 13:58:25.289341383 +0000 UTC m=+148.125134240" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.317672 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cl6js" event={"ID":"2daa96a6-fc64-4fd3-ba1a-d4ae6bbf97b0","Type":"ContainerStarted","Data":"946505ba43511a36ad5ec2f92313a154b5272751641cf3d317eac966cd7d4051"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.333926 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tvwsv" podStartSLOduration=129.333905096 podStartE2EDuration="2m9.333905096s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:25.333874405 +0000 UTC m=+148.169667262" watchObservedRunningTime="2025-12-06 13:58:25.333905096 +0000 UTC m=+148.169697953" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.351391 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dp75h" event={"ID":"806ad18b-e408-4679-9a37-b6efd48dd4d8","Type":"ContainerStarted","Data":"4f9c996d4bb4706af9946a955e367538dd915fd8ea361a0d56a9c3c6d77ad290"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.358458 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-m9fbd" event={"ID":"38703ec7-616b-499d-a173-4ea254e11d9a","Type":"ContainerStarted","Data":"e70efd49f2e919662b6c63df78d2a7941adc63a28938d943f90f7de429a59c90"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.393707 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:25 crc kubenswrapper[4706]: E1206 13:58:25.397969 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:25.897943253 +0000 UTC m=+148.733736110 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.430458 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4" podStartSLOduration=128.424365147 podStartE2EDuration="2m8.424365147s" podCreationTimestamp="2025-12-06 13:56:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:25.423062054 +0000 UTC m=+148.258854911" watchObservedRunningTime="2025-12-06 13:58:25.424365147 +0000 UTC m=+148.260158004" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.430839 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ldd62" event={"ID":"532ce376-db1f-4770-a4cd-a30318242daf","Type":"ContainerStarted","Data":"13271b153a44d4aea51ac2774758878dab6de9a0e31a0a34b2ba061c5c187167"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.430877 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ldd62" event={"ID":"532ce376-db1f-4770-a4cd-a30318242daf","Type":"ContainerStarted","Data":"777d516f264e4e1e039085210b452e734d6c83f56f7675e4f5c94e466fc7d5ea"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.431343 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" podStartSLOduration=128.431335478 podStartE2EDuration="2m8.431335478s" podCreationTimestamp="2025-12-06 13:56:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:25.373226584 +0000 UTC m=+148.209019441" watchObservedRunningTime="2025-12-06 13:58:25.431335478 +0000 UTC m=+148.267128335" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.440501 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" event={"ID":"fe23e3f5-57af-4fd2-a8bd-9364bbe39776","Type":"ContainerStarted","Data":"3b9872c83d17ddc383464b884a220cdaff31cf0a9590816d2ba8d02334a34d8d"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.478334 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-ldd62" podStartSLOduration=129.478319764 podStartE2EDuration="2m9.478319764s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:25.477723738 +0000 UTC m=+148.313516595" watchObservedRunningTime="2025-12-06 13:58:25.478319764 +0000 UTC m=+148.314112621" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.504519 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-j4chj" podStartSLOduration=129.504504111 podStartE2EDuration="2m9.504504111s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:25.502706905 +0000 UTC m=+148.338499762" watchObservedRunningTime="2025-12-06 13:58:25.504504111 +0000 UTC m=+148.340296958" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.508798 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:25 crc kubenswrapper[4706]: E1206 13:58:25.509102 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:26.00908905 +0000 UTC m=+148.844881907 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.526646 4706 patch_prober.go:28] interesting pod/downloads-7954f5f757-9jn7l container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.536908 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9jn7l" podUID="621a48da-c965-4d74-abd0-62e119fd81c8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.587153 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.587178 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" event={"ID":"f6f43404-a6e1-49ee-bbef-9e027779cef5","Type":"ContainerStarted","Data":"326ba42b778fe64a6dd7094dede87758b89f40146124b4b393beda60b79631c5"} Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.587198 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.609567 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:25 crc kubenswrapper[4706]: E1206 13:58:25.611579 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:26.111557982 +0000 UTC m=+148.947350839 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.619029 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.619239 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:25 crc kubenswrapper[4706]: E1206 13:58:25.619786 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:26.119771554 +0000 UTC m=+148.955564411 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.651842 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" podStartSLOduration=129.651827684 podStartE2EDuration="2m9.651827684s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:25.571879065 +0000 UTC m=+148.407671922" watchObservedRunningTime="2025-12-06 13:58:25.651827684 +0000 UTC m=+148.487620541" Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.720644 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:25 crc kubenswrapper[4706]: E1206 13:58:25.721739 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:26.221723573 +0000 UTC m=+149.057516430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.835850 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:25 crc kubenswrapper[4706]: E1206 13:58:25.836174 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:26.336160004 +0000 UTC m=+149.171952861 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.938504 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:25 crc kubenswrapper[4706]: E1206 13:58:25.938696 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:26.438660617 +0000 UTC m=+149.274453474 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.938866 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:25 crc kubenswrapper[4706]: E1206 13:58:25.939452 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:26.439432387 +0000 UTC m=+149.275225254 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.947614 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:25 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:25 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:25 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:25 crc kubenswrapper[4706]: I1206 13:58:25.947678 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.043194 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:26 crc kubenswrapper[4706]: E1206 13:58:26.044066 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:26.544039184 +0000 UTC m=+149.379832041 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.146073 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:26 crc kubenswrapper[4706]: E1206 13:58:26.146459 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:26.646443684 +0000 UTC m=+149.482236541 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.253353 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:26 crc kubenswrapper[4706]: E1206 13:58:26.254469 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:26.754444359 +0000 UTC m=+149.590237216 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.360571 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:26 crc kubenswrapper[4706]: E1206 13:58:26.360924 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:26.860912174 +0000 UTC m=+149.696705031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:26 crc kubenswrapper[4706]: W1206 13:58:26.372459 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-1c3842c1becbcbcb809dd3fef4943d57422f8803f32279261f9f681cda478b5a WatchSource:0}: Error finding container 1c3842c1becbcbcb809dd3fef4943d57422f8803f32279261f9f681cda478b5a: Status 404 returned error can't find the container with id 1c3842c1becbcbcb809dd3fef4943d57422f8803f32279261f9f681cda478b5a Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.472363 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:26 crc kubenswrapper[4706]: E1206 13:58:26.472934 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:26.972907122 +0000 UTC m=+149.808699979 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.473425 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:26 crc kubenswrapper[4706]: E1206 13:58:26.473942 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:26.973922649 +0000 UTC m=+149.809715506 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.577476 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:26 crc kubenswrapper[4706]: E1206 13:58:26.577824 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:27.077808107 +0000 UTC m=+149.913600964 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.653012 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-zx8fv" event={"ID":"1a8d709a-565d-4213-ae67-17440c2d7678","Type":"ContainerStarted","Data":"d3534e482b9f681e6b9a4af7a92db12d9bda0309862ea918879fda0c381d2d98"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.673498 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" event={"ID":"d0d1509c-c26c-4066-8dab-23d0205d4211","Type":"ContainerStarted","Data":"60e2ad816ebfacf030054af69b320e00ca07bbff8168234939e41f4f3746cef2"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.673952 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" event={"ID":"d0d1509c-c26c-4066-8dab-23d0205d4211","Type":"ContainerStarted","Data":"83436ec78e870ce4935e1992947143a709a092a7c40f6a30bdf2fb97d50d4a9e"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.679503 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:26 crc kubenswrapper[4706]: E1206 13:58:26.679766 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:27.179755166 +0000 UTC m=+150.015548023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.713132 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-m9fbd" event={"ID":"38703ec7-616b-499d-a173-4ea254e11d9a","Type":"ContainerStarted","Data":"1559df6884a6abebfba76e2843480f32aa9b4f6d53075ff5797ecc78b957e1ae"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.726639 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2qmsq" podStartSLOduration=130.726617658 podStartE2EDuration="2m10.726617658s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:26.724884004 +0000 UTC m=+149.560676861" watchObservedRunningTime="2025-12-06 13:58:26.726617658 +0000 UTC m=+149.562410515" Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.734045 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-zx8fv" podStartSLOduration=130.73403085 podStartE2EDuration="2m10.73403085s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:26.67878419 +0000 UTC m=+149.514577047" watchObservedRunningTime="2025-12-06 13:58:26.73403085 +0000 UTC m=+149.569823707" Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.763671 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2t78" event={"ID":"d2b3bc37-36b2-473e-ae10-b0bdc30f2ef2","Type":"ContainerStarted","Data":"4935f91a961614815ba88c53af93d9f2d029094e6b22f9b6756b1f464da55d24"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.772529 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ebd569995ea9b48104afe2edece459fe316fe1afe673ee115515d0d340486153"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.785566 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:26 crc kubenswrapper[4706]: E1206 13:58:26.785812 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:27.285779049 +0000 UTC m=+150.121571906 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.785913 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:26 crc kubenswrapper[4706]: E1206 13:58:26.787489 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:27.287471293 +0000 UTC m=+150.123264150 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.796020 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-97565" event={"ID":"65a55c43-6a12-46dc-b181-8241ceab5d12","Type":"ContainerStarted","Data":"7676c830fa5a75ca5e6ed86a420c70e4a1269d774a87882044233ccc5bdc0d03"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.807464 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" event={"ID":"09f7b3f2-ef36-4a21-970a-3a19b7a6a726","Type":"ContainerStarted","Data":"d47f8ab70bc2611e3c0ed54bcbf967880cd71bcd2716ed2593aa54de641aedbe"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.818607 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-r2t78" podStartSLOduration=130.818585019 podStartE2EDuration="2m10.818585019s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:26.797535974 +0000 UTC m=+149.633328831" watchObservedRunningTime="2025-12-06 13:58:26.818585019 +0000 UTC m=+149.654377876" Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.821360 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-97565" podStartSLOduration=130.82133984 podStartE2EDuration="2m10.82133984s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:26.816248008 +0000 UTC m=+149.652040865" watchObservedRunningTime="2025-12-06 13:58:26.82133984 +0000 UTC m=+149.657132697" Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.843517 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-t2255" event={"ID":"a0b36426-e051-4864-b98e-b1ecf89914a4","Type":"ContainerStarted","Data":"bf61eca629bbf7111afde4ab6dccd723a2b1a399fd8622bee50686aa7c36c751"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.884906 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dp75h" event={"ID":"806ad18b-e408-4679-9a37-b6efd48dd4d8","Type":"ContainerStarted","Data":"8e36f29cc1e54007a0e2fbb51d66419fd909d549314276d4daf6bd59ea673d14"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.884954 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dp75h" event={"ID":"806ad18b-e408-4679-9a37-b6efd48dd4d8","Type":"ContainerStarted","Data":"227672846b86076779f562b16be31755f8928bf861357e9f482de25a4e13e153"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.885525 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dp75h" Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.887333 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:26 crc kubenswrapper[4706]: E1206 13:58:26.888333 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:27.388300153 +0000 UTC m=+150.224093010 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.889721 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-t2255" Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.916331 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dp75h" podStartSLOduration=129.916302087 podStartE2EDuration="2m9.916302087s" podCreationTimestamp="2025-12-06 13:56:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:26.914050809 +0000 UTC m=+149.749843666" watchObservedRunningTime="2025-12-06 13:58:26.916302087 +0000 UTC m=+149.752094944" Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.921069 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56" event={"ID":"1a50e9ce-9db6-4f82-b1a6-6cf0e1b97d90","Type":"ContainerStarted","Data":"fd0d861e6b1fd8f168507ddcb6cece01c18e5573e191b29100e5c6a3bacf5569"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.926454 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d5zwg" Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.933206 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx" event={"ID":"1ae74e67-c9db-4ba6-87e5-0fcb923244b3","Type":"ContainerStarted","Data":"13d949b60ea3ea8732f99c55e27a5ce1b69953badfeec5343a71a01b63230fe3"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.933253 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx" event={"ID":"1ae74e67-c9db-4ba6-87e5-0fcb923244b3","Type":"ContainerStarted","Data":"43bc6dd3306e473d6c1909af8307feb307762bdbc32d814990a50e84715829e4"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.944392 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:26 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:26 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:26 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.944985 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.956291 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"6e6bac8c99249245ef65be38fb2420c82f31b3f04fa4dff7dfa5ba33096228dc"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.958017 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kdfk5" event={"ID":"407db426-50cd-413d-bd5a-392aed7b1897","Type":"ContainerStarted","Data":"54aac910f2338e745d7218bdac215593d7e42bc1972f70a0a48d07ffec5c0f8b"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.959070 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cl6js" event={"ID":"2daa96a6-fc64-4fd3-ba1a-d4ae6bbf97b0","Type":"ContainerStarted","Data":"eb12462812f1e556bb73548de8bc091a14ab641a2ad2ff0979153a57402511ee"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.994670 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" event={"ID":"7898190a-5957-4772-abbb-1bfbf16f4eb5","Type":"ContainerStarted","Data":"4b0d06d22caa9b797f856c01a75d8a5de765373ca71b47a6c4c24cff7402ae88"} Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.995325 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:26 crc kubenswrapper[4706]: I1206 13:58:26.995966 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:26 crc kubenswrapper[4706]: E1206 13:58:26.997721 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:27.497709504 +0000 UTC m=+150.333502361 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.006201 4706 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-l55q6 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.18:6443/healthz\": dial tcp 10.217.0.18:6443: connect: connection refused" start-of-body= Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.006248 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" podUID="7898190a-5957-4772-abbb-1bfbf16f4eb5" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.18:6443/healthz\": dial tcp 10.217.0.18:6443: connect: connection refused" Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.033046 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx2jx" podStartSLOduration=131.033012328 podStartE2EDuration="2m11.033012328s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:27.029810555 +0000 UTC m=+149.865603412" watchObservedRunningTime="2025-12-06 13:58:27.033012328 +0000 UTC m=+149.868805185" Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.033543 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bkh56" podStartSLOduration=131.033535871 podStartE2EDuration="2m11.033535871s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:26.99485805 +0000 UTC m=+149.830650927" watchObservedRunningTime="2025-12-06 13:58:27.033535871 +0000 UTC m=+149.869328728" Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.035101 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rpnhr" event={"ID":"f6f43404-a6e1-49ee-bbef-9e027779cef5","Type":"ContainerStarted","Data":"11f55350d517c520d91df7ff970f33e894454bc5155d42a48bfca63692fa8393"} Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.057576 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs" event={"ID":"aca05689-5cad-4e13-b513-1ef9de804ba9","Type":"ContainerStarted","Data":"0780b0b05b7b805f43143e361e1c773d9712a32930b94ea45e9c3d4f6b2f40ff"} Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.058492 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs" Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.070156 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"1c3842c1becbcbcb809dd3fef4943d57422f8803f32279261f9f681cda478b5a"} Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.070744 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.080934 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs" Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.091048 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4" event={"ID":"f65fc54c-f0ca-40dc-a8bc-9084483d2d11","Type":"ContainerStarted","Data":"b34edd6d0bc8c08d07d16108592e046267f10291823c86a72200b99b09cb7f7c"} Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.100634 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.101737 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-kdfk5" podStartSLOduration=8.101723036 podStartE2EDuration="8.101723036s" podCreationTimestamp="2025-12-06 13:58:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:27.100184116 +0000 UTC m=+149.935976973" watchObservedRunningTime="2025-12-06 13:58:27.101723036 +0000 UTC m=+149.937515893" Dec 06 13:58:27 crc kubenswrapper[4706]: E1206 13:58:27.101823 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:27.601808898 +0000 UTC m=+150.437601755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.117227 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" event={"ID":"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3","Type":"ContainerStarted","Data":"c2a9dfcb1b08640b2730c92f11f465c17533636004a95e336dbcb9bbe7725507"} Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.118380 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfkh4" Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.119649 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.125138 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-cl6js" podStartSLOduration=131.125125662 podStartE2EDuration="2m11.125125662s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:27.124723531 +0000 UTC m=+149.960516388" watchObservedRunningTime="2025-12-06 13:58:27.125125662 +0000 UTC m=+149.960918509" Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.125512 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.133063 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.133091 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2wmlg" Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.133125 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfd8q" Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.191114 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mfhhs" podStartSLOduration=130.191091489 podStartE2EDuration="2m10.191091489s" podCreationTimestamp="2025-12-06 13:56:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:27.189485397 +0000 UTC m=+150.025278254" watchObservedRunningTime="2025-12-06 13:58:27.191091489 +0000 UTC m=+150.026884346" Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.209837 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:27 crc kubenswrapper[4706]: E1206 13:58:27.212746 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:27.712734799 +0000 UTC m=+150.548527646 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.231985 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" podStartSLOduration=131.231968347 podStartE2EDuration="2m11.231968347s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:27.223950969 +0000 UTC m=+150.059743836" watchObservedRunningTime="2025-12-06 13:58:27.231968347 +0000 UTC m=+150.067761214" Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.285831 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" podStartSLOduration=131.28581362 podStartE2EDuration="2m11.28581362s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:27.284397814 +0000 UTC m=+150.120190671" watchObservedRunningTime="2025-12-06 13:58:27.28581362 +0000 UTC m=+150.121606477" Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.311014 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:27 crc kubenswrapper[4706]: E1206 13:58:27.311585 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:27.811558406 +0000 UTC m=+150.647351263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.312600 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:27 crc kubenswrapper[4706]: E1206 13:58:27.319775 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:27.819760129 +0000 UTC m=+150.655552986 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.414841 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:27 crc kubenswrapper[4706]: E1206 13:58:27.415191 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:27.915174008 +0000 UTC m=+150.750966865 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.518894 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:27 crc kubenswrapper[4706]: E1206 13:58:27.519718 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:28.019698103 +0000 UTC m=+150.855490960 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.622570 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:27 crc kubenswrapper[4706]: E1206 13:58:27.622854 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:28.122836142 +0000 UTC m=+150.958628999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.723142 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:27 crc kubenswrapper[4706]: E1206 13:58:27.723529 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:28.223513678 +0000 UTC m=+151.059306535 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.826997 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:27 crc kubenswrapper[4706]: E1206 13:58:27.828178 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:28.328138915 +0000 UTC m=+151.163931772 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.928714 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:27 crc kubenswrapper[4706]: E1206 13:58:27.929096 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:28.429080268 +0000 UTC m=+151.264873125 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.943062 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:27 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:27 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:27 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:27 crc kubenswrapper[4706]: I1206 13:58:27.943190 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.029448 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:28 crc kubenswrapper[4706]: E1206 13:58:28.029677 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:28.52964056 +0000 UTC m=+151.365433417 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.029948 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:28 crc kubenswrapper[4706]: E1206 13:58:28.030215 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:28.530202285 +0000 UTC m=+151.365995142 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.059555 4706 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.122024 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cl6js" event={"ID":"2daa96a6-fc64-4fd3-ba1a-d4ae6bbf97b0","Type":"ContainerStarted","Data":"64f180a85611ed5b1743df8ab681c92458042f2f7227214679ba925c90ee363e"} Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.124392 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" event={"ID":"0acf2c47-e9d7-4759-aa80-9f4ca8488899","Type":"ContainerStarted","Data":"f271c3827f8033e423548f197b37e981ac636a817458a85ad88b2ff5fd4a914e"} Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.125741 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" event={"ID":"27733066-fbcd-474e-9af2-d2bd168533c4","Type":"ContainerStarted","Data":"80fbc84a05e1ed21add35ff9b999315ac9dfee7675e6e749f99b0d1a53c1c18f"} Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.125762 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" event={"ID":"27733066-fbcd-474e-9af2-d2bd168533c4","Type":"ContainerStarted","Data":"89fee37664046a1ea90b3eb37231de5169c1fd1c05c542794bc17c0df49bd5d2"} Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.126870 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-m9fbd" event={"ID":"38703ec7-616b-499d-a173-4ea254e11d9a","Type":"ContainerStarted","Data":"21962741d841cb75d4e981503560f4deb3184a6c4d0907c41f8a24c1970c4709"} Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.127204 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-m9fbd" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.128958 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f0d44bdaf8c9c6395190742462a7cf6aa95399d87dec1c859c7ca2b46d2964f8"} Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.130203 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"faa5dab14dd4bf1738fd49a57b683b8d0c4f8d3d0b34ca7201b092e042d025bf"} Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.130990 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:28 crc kubenswrapper[4706]: E1206 13:58:28.131071 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:28.631051915 +0000 UTC m=+151.466844772 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.131502 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.131693 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"eae88a69304a2cf8e3bb3459fb2f12376d3195ebbd4596cb8a2e03b3cb96603f"} Dec 06 13:58:28 crc kubenswrapper[4706]: E1206 13:58:28.132027 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:28.63201025 +0000 UTC m=+151.467803107 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.200983 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" podStartSLOduration=132.200949294 podStartE2EDuration="2m12.200949294s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:28.164527261 +0000 UTC m=+151.000320118" watchObservedRunningTime="2025-12-06 13:58:28.200949294 +0000 UTC m=+151.036742151" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.201735 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-m9fbd" podStartSLOduration=9.201729414 podStartE2EDuration="9.201729414s" podCreationTimestamp="2025-12-06 13:58:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:28.199373353 +0000 UTC m=+151.035166210" watchObservedRunningTime="2025-12-06 13:58:28.201729414 +0000 UTC m=+151.037522281" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.232851 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:28 crc kubenswrapper[4706]: E1206 13:58:28.233018 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:28.732989813 +0000 UTC m=+151.568782670 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.233593 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:28 crc kubenswrapper[4706]: E1206 13:58:28.237602 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:28.737590832 +0000 UTC m=+151.573383689 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.335279 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:28 crc kubenswrapper[4706]: E1206 13:58:28.336138 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:28.836122172 +0000 UTC m=+151.671915029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.351558 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hjkcq"] Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.364720 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hjkcq" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.369969 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.374664 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hjkcq"] Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.437143 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:28 crc kubenswrapper[4706]: E1206 13:58:28.437798 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:28.937782343 +0000 UTC m=+151.773575200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.538489 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:28 crc kubenswrapper[4706]: E1206 13:58:28.538626 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:29.038605562 +0000 UTC m=+151.874398419 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.538702 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.538743 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-utilities\") pod \"community-operators-hjkcq\" (UID: \"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5\") " pod="openshift-marketplace/community-operators-hjkcq" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.538764 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-catalog-content\") pod \"community-operators-hjkcq\" (UID: \"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5\") " pod="openshift-marketplace/community-operators-hjkcq" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.538783 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxrzw\" (UniqueName: \"kubernetes.io/projected/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-kube-api-access-fxrzw\") pod \"community-operators-hjkcq\" (UID: \"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5\") " pod="openshift-marketplace/community-operators-hjkcq" Dec 06 13:58:28 crc kubenswrapper[4706]: E1206 13:58:28.539067 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:29.039055674 +0000 UTC m=+151.874848531 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.554022 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kslt9"] Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.555077 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kslt9" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.558776 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.564916 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kslt9"] Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.640133 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:28 crc kubenswrapper[4706]: E1206 13:58:28.640346 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:29.140301804 +0000 UTC m=+151.976094661 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.640388 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-utilities\") pod \"community-operators-hjkcq\" (UID: \"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5\") " pod="openshift-marketplace/community-operators-hjkcq" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.640418 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-catalog-content\") pod \"community-operators-hjkcq\" (UID: \"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5\") " pod="openshift-marketplace/community-operators-hjkcq" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.640437 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxrzw\" (UniqueName: \"kubernetes.io/projected/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-kube-api-access-fxrzw\") pod \"community-operators-hjkcq\" (UID: \"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5\") " pod="openshift-marketplace/community-operators-hjkcq" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.640506 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:28 crc kubenswrapper[4706]: E1206 13:58:28.640782 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 13:58:29.140771006 +0000 UTC m=+151.976563863 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4bwl" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.640815 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-utilities\") pod \"community-operators-hjkcq\" (UID: \"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5\") " pod="openshift-marketplace/community-operators-hjkcq" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.641155 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-catalog-content\") pod \"community-operators-hjkcq\" (UID: \"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5\") " pod="openshift-marketplace/community-operators-hjkcq" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.662228 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxrzw\" (UniqueName: \"kubernetes.io/projected/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-kube-api-access-fxrzw\") pod \"community-operators-hjkcq\" (UID: \"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5\") " pod="openshift-marketplace/community-operators-hjkcq" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.662868 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.721432 4706 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-06T13:58:28.059590015Z","Handler":null,"Name":""} Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.733254 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hjkcq" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.741491 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.741677 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-catalog-content\") pod \"certified-operators-kslt9\" (UID: \"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2\") " pod="openshift-marketplace/certified-operators-kslt9" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.741726 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p4zf\" (UniqueName: \"kubernetes.io/projected/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-kube-api-access-6p4zf\") pod \"certified-operators-kslt9\" (UID: \"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2\") " pod="openshift-marketplace/certified-operators-kslt9" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.741799 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-utilities\") pod \"certified-operators-kslt9\" (UID: \"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2\") " pod="openshift-marketplace/certified-operators-kslt9" Dec 06 13:58:28 crc kubenswrapper[4706]: E1206 13:58:28.741941 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 13:58:29.241922614 +0000 UTC m=+152.077715471 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.754739 4706 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.754783 4706 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.754751 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zhlvc"] Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.755907 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zhlvc" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.774628 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zhlvc"] Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.842966 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.843019 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-catalog-content\") pod \"certified-operators-kslt9\" (UID: \"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2\") " pod="openshift-marketplace/certified-operators-kslt9" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.843048 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p4zf\" (UniqueName: \"kubernetes.io/projected/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-kube-api-access-6p4zf\") pod \"certified-operators-kslt9\" (UID: \"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2\") " pod="openshift-marketplace/certified-operators-kslt9" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.843122 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-utilities\") pod \"certified-operators-kslt9\" (UID: \"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2\") " pod="openshift-marketplace/certified-operators-kslt9" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.843538 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-utilities\") pod \"certified-operators-kslt9\" (UID: \"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2\") " pod="openshift-marketplace/certified-operators-kslt9" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.843855 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-catalog-content\") pod \"certified-operators-kslt9\" (UID: \"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2\") " pod="openshift-marketplace/certified-operators-kslt9" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.848760 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.848799 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.864216 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p4zf\" (UniqueName: \"kubernetes.io/projected/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-kube-api-access-6p4zf\") pod \"certified-operators-kslt9\" (UID: \"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2\") " pod="openshift-marketplace/certified-operators-kslt9" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.885496 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kslt9" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.889290 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4bwl\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.933700 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:28 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:28 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:28 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.933778 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.944965 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.946760 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d12f4107-0d89-4d3c-958f-7355fbd27575-utilities\") pod \"community-operators-zhlvc\" (UID: \"d12f4107-0d89-4d3c-958f-7355fbd27575\") " pod="openshift-marketplace/community-operators-zhlvc" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.946823 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w4bj\" (UniqueName: \"kubernetes.io/projected/d12f4107-0d89-4d3c-958f-7355fbd27575-kube-api-access-9w4bj\") pod \"community-operators-zhlvc\" (UID: \"d12f4107-0d89-4d3c-958f-7355fbd27575\") " pod="openshift-marketplace/community-operators-zhlvc" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.946905 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d12f4107-0d89-4d3c-958f-7355fbd27575-catalog-content\") pod \"community-operators-zhlvc\" (UID: \"d12f4107-0d89-4d3c-958f-7355fbd27575\") " pod="openshift-marketplace/community-operators-zhlvc" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.976002 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gx45h"] Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.976546 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.977041 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gx45h" Dec 06 13:58:28 crc kubenswrapper[4706]: I1206 13:58:28.997984 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gx45h"] Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.048915 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d12f4107-0d89-4d3c-958f-7355fbd27575-catalog-content\") pod \"community-operators-zhlvc\" (UID: \"d12f4107-0d89-4d3c-958f-7355fbd27575\") " pod="openshift-marketplace/community-operators-zhlvc" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.048980 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d12f4107-0d89-4d3c-958f-7355fbd27575-utilities\") pod \"community-operators-zhlvc\" (UID: \"d12f4107-0d89-4d3c-958f-7355fbd27575\") " pod="openshift-marketplace/community-operators-zhlvc" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.049004 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w4bj\" (UniqueName: \"kubernetes.io/projected/d12f4107-0d89-4d3c-958f-7355fbd27575-kube-api-access-9w4bj\") pod \"community-operators-zhlvc\" (UID: \"d12f4107-0d89-4d3c-958f-7355fbd27575\") " pod="openshift-marketplace/community-operators-zhlvc" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.050477 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d12f4107-0d89-4d3c-958f-7355fbd27575-catalog-content\") pod \"community-operators-zhlvc\" (UID: \"d12f4107-0d89-4d3c-958f-7355fbd27575\") " pod="openshift-marketplace/community-operators-zhlvc" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.053606 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d12f4107-0d89-4d3c-958f-7355fbd27575-utilities\") pod \"community-operators-zhlvc\" (UID: \"d12f4107-0d89-4d3c-958f-7355fbd27575\") " pod="openshift-marketplace/community-operators-zhlvc" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.067455 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w4bj\" (UniqueName: \"kubernetes.io/projected/d12f4107-0d89-4d3c-958f-7355fbd27575-kube-api-access-9w4bj\") pod \"community-operators-zhlvc\" (UID: \"d12f4107-0d89-4d3c-958f-7355fbd27575\") " pod="openshift-marketplace/community-operators-zhlvc" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.114073 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zhlvc" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.149872 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" event={"ID":"27733066-fbcd-474e-9af2-d2bd168533c4","Type":"ContainerStarted","Data":"eccc41d11c534e156478eeb5d26e96515c21a111a48016812f4b1f268ff0d4f3"} Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.149945 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" event={"ID":"27733066-fbcd-474e-9af2-d2bd168533c4","Type":"ContainerStarted","Data":"9f4b0880e5e203865b6f49a1c91dc264a806bbbaa7ee7f2a7650170e91f6877d"} Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.151378 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l68w\" (UniqueName: \"kubernetes.io/projected/06e9400a-9432-4f75-af53-78b3b3facd7a-kube-api-access-2l68w\") pod \"certified-operators-gx45h\" (UID: \"06e9400a-9432-4f75-af53-78b3b3facd7a\") " pod="openshift-marketplace/certified-operators-gx45h" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.151417 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06e9400a-9432-4f75-af53-78b3b3facd7a-utilities\") pod \"certified-operators-gx45h\" (UID: \"06e9400a-9432-4f75-af53-78b3b3facd7a\") " pod="openshift-marketplace/certified-operators-gx45h" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.151523 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06e9400a-9432-4f75-af53-78b3b3facd7a-catalog-content\") pod \"certified-operators-gx45h\" (UID: \"06e9400a-9432-4f75-af53-78b3b3facd7a\") " pod="openshift-marketplace/certified-operators-gx45h" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.153020 4706 generic.go:334] "Generic (PLEG): container finished" podID="ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3" containerID="c2a9dfcb1b08640b2730c92f11f465c17533636004a95e336dbcb9bbe7725507" exitCode=0 Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.157446 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" event={"ID":"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3","Type":"ContainerDied","Data":"c2a9dfcb1b08640b2730c92f11f465c17533636004a95e336dbcb9bbe7725507"} Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.166483 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.212192 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-7xn8m" podStartSLOduration=10.212128323 podStartE2EDuration="10.212128323s" podCreationTimestamp="2025-12-06 13:58:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:29.190331499 +0000 UTC m=+152.026124356" watchObservedRunningTime="2025-12-06 13:58:29.212128323 +0000 UTC m=+152.047921170" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.219389 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kslt9"] Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.253574 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06e9400a-9432-4f75-af53-78b3b3facd7a-catalog-content\") pod \"certified-operators-gx45h\" (UID: \"06e9400a-9432-4f75-af53-78b3b3facd7a\") " pod="openshift-marketplace/certified-operators-gx45h" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.253852 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l68w\" (UniqueName: \"kubernetes.io/projected/06e9400a-9432-4f75-af53-78b3b3facd7a-kube-api-access-2l68w\") pod \"certified-operators-gx45h\" (UID: \"06e9400a-9432-4f75-af53-78b3b3facd7a\") " pod="openshift-marketplace/certified-operators-gx45h" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.253903 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06e9400a-9432-4f75-af53-78b3b3facd7a-utilities\") pod \"certified-operators-gx45h\" (UID: \"06e9400a-9432-4f75-af53-78b3b3facd7a\") " pod="openshift-marketplace/certified-operators-gx45h" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.255115 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06e9400a-9432-4f75-af53-78b3b3facd7a-utilities\") pod \"certified-operators-gx45h\" (UID: \"06e9400a-9432-4f75-af53-78b3b3facd7a\") " pod="openshift-marketplace/certified-operators-gx45h" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.257147 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06e9400a-9432-4f75-af53-78b3b3facd7a-catalog-content\") pod \"certified-operators-gx45h\" (UID: \"06e9400a-9432-4f75-af53-78b3b3facd7a\") " pod="openshift-marketplace/certified-operators-gx45h" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.282448 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hjkcq"] Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.291215 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l68w\" (UniqueName: \"kubernetes.io/projected/06e9400a-9432-4f75-af53-78b3b3facd7a-kube-api-access-2l68w\") pod \"certified-operators-gx45h\" (UID: \"06e9400a-9432-4f75-af53-78b3b3facd7a\") " pod="openshift-marketplace/certified-operators-gx45h" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.303356 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gx45h" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.376334 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zhlvc"] Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.489348 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-z4bwl"] Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.518369 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.596060 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gx45h"] Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.616488 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.617102 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.619497 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.619648 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.623479 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 13:58:29 crc kubenswrapper[4706]: W1206 13:58:29.696597 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06e9400a_9432_4f75_af53_78b3b3facd7a.slice/crio-cb2829434fb2137e63a72f6f9fde965d2074d9bb96386957b60a6d708a2704d2 WatchSource:0}: Error finding container cb2829434fb2137e63a72f6f9fde965d2074d9bb96386957b60a6d708a2704d2: Status 404 returned error can't find the container with id cb2829434fb2137e63a72f6f9fde965d2074d9bb96386957b60a6d708a2704d2 Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.782557 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3904e406-8755-45fa-8088-93ecca405707-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3904e406-8755-45fa-8088-93ecca405707\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.782615 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3904e406-8755-45fa-8088-93ecca405707-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3904e406-8755-45fa-8088-93ecca405707\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.884869 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3904e406-8755-45fa-8088-93ecca405707-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3904e406-8755-45fa-8088-93ecca405707\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.884936 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3904e406-8755-45fa-8088-93ecca405707-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3904e406-8755-45fa-8088-93ecca405707\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.885082 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3904e406-8755-45fa-8088-93ecca405707-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3904e406-8755-45fa-8088-93ecca405707\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.907734 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3904e406-8755-45fa-8088-93ecca405707-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3904e406-8755-45fa-8088-93ecca405707\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.935014 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:29 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:29 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:29 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:29 crc kubenswrapper[4706]: I1206 13:58:29.935081 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.006600 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.160103 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" event={"ID":"ed999748-8e6d-4402-81e3-1f483b8440e8","Type":"ContainerStarted","Data":"164ee49b83f94279c8df627e642fa48664a0008fed4c3c0d55d6a6a729a656f7"} Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.160700 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" event={"ID":"ed999748-8e6d-4402-81e3-1f483b8440e8","Type":"ContainerStarted","Data":"de0693966858e44e25daaeb7d8a9be5818024e1c46f58dabd14c84a6ee4e9fe5"} Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.160746 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.162588 4706 generic.go:334] "Generic (PLEG): container finished" podID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" containerID="aac2689774446cc2f9ba1997a613d4e735786174daa0e2e0135c7f73aaeeb65d" exitCode=0 Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.162659 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjkcq" event={"ID":"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5","Type":"ContainerDied","Data":"aac2689774446cc2f9ba1997a613d4e735786174daa0e2e0135c7f73aaeeb65d"} Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.162685 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjkcq" event={"ID":"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5","Type":"ContainerStarted","Data":"badef2b00d9a2dc3a88364f683c49374cd32ae3dd24770c54b8f16b457a3c0b0"} Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.165639 4706 generic.go:334] "Generic (PLEG): container finished" podID="d12f4107-0d89-4d3c-958f-7355fbd27575" containerID="6447594ac6b38ab5e2a58a1551c4b2243c7ae35e0aae2468cb919993d61123c4" exitCode=0 Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.165754 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zhlvc" event={"ID":"d12f4107-0d89-4d3c-958f-7355fbd27575","Type":"ContainerDied","Data":"6447594ac6b38ab5e2a58a1551c4b2243c7ae35e0aae2468cb919993d61123c4"} Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.165800 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zhlvc" event={"ID":"d12f4107-0d89-4d3c-958f-7355fbd27575","Type":"ContainerStarted","Data":"79e0531c01ab0055b715bce21638bcba539340cee7ae422b15aff937e9f69799"} Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.166910 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.168230 4706 generic.go:334] "Generic (PLEG): container finished" podID="06e9400a-9432-4f75-af53-78b3b3facd7a" containerID="114116940b690d4be2b7e173f264304b324950eb3a9911e95a70f5f8e45b8dfc" exitCode=0 Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.168326 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gx45h" event={"ID":"06e9400a-9432-4f75-af53-78b3b3facd7a","Type":"ContainerDied","Data":"114116940b690d4be2b7e173f264304b324950eb3a9911e95a70f5f8e45b8dfc"} Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.168363 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gx45h" event={"ID":"06e9400a-9432-4f75-af53-78b3b3facd7a","Type":"ContainerStarted","Data":"cb2829434fb2137e63a72f6f9fde965d2074d9bb96386957b60a6d708a2704d2"} Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.176843 4706 generic.go:334] "Generic (PLEG): container finished" podID="1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" containerID="7f974c473ff680877f09f18406c1f699e8c3f551fd1bafb7cada66e04cb0bfa2" exitCode=0 Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.177215 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kslt9" event={"ID":"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2","Type":"ContainerDied","Data":"7f974c473ff680877f09f18406c1f699e8c3f551fd1bafb7cada66e04cb0bfa2"} Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.177788 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kslt9" event={"ID":"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2","Type":"ContainerStarted","Data":"dcfdb05768a03816b2f74b6417276cb768e3940493e24a62398056ff02337b6d"} Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.193142 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" podStartSLOduration=134.19311557 podStartE2EDuration="2m14.19311557s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:30.192619467 +0000 UTC m=+153.028412384" watchObservedRunningTime="2025-12-06 13:58:30.19311557 +0000 UTC m=+153.028908467" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.261050 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 13:58:30 crc kubenswrapper[4706]: W1206 13:58:30.289694 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod3904e406_8755_45fa_8088_93ecca405707.slice/crio-d87a0436101175c7a6d8e7dbdddcb1d1a876f3d7feb601668699c456b89eb1c6 WatchSource:0}: Error finding container d87a0436101175c7a6d8e7dbdddcb1d1a876f3d7feb601668699c456b89eb1c6: Status 404 returned error can't find the container with id d87a0436101175c7a6d8e7dbdddcb1d1a876f3d7feb601668699c456b89eb1c6 Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.360090 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wdw9k"] Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.361665 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdw9k" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.363755 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.365223 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdw9k"] Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.423342 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.497265 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbp2v\" (UniqueName: \"kubernetes.io/projected/906c227f-a713-49d2-acfd-633646b6e0da-kube-api-access-qbp2v\") pod \"redhat-marketplace-wdw9k\" (UID: \"906c227f-a713-49d2-acfd-633646b6e0da\") " pod="openshift-marketplace/redhat-marketplace-wdw9k" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.497330 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906c227f-a713-49d2-acfd-633646b6e0da-utilities\") pod \"redhat-marketplace-wdw9k\" (UID: \"906c227f-a713-49d2-acfd-633646b6e0da\") " pod="openshift-marketplace/redhat-marketplace-wdw9k" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.497364 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906c227f-a713-49d2-acfd-633646b6e0da-catalog-content\") pod \"redhat-marketplace-wdw9k\" (UID: \"906c227f-a713-49d2-acfd-633646b6e0da\") " pod="openshift-marketplace/redhat-marketplace-wdw9k" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.598261 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6b88\" (UniqueName: \"kubernetes.io/projected/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-kube-api-access-g6b88\") pod \"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3\" (UID: \"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3\") " Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.598350 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-config-volume\") pod \"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3\" (UID: \"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3\") " Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.598439 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-secret-volume\") pod \"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3\" (UID: \"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3\") " Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.598624 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbp2v\" (UniqueName: \"kubernetes.io/projected/906c227f-a713-49d2-acfd-633646b6e0da-kube-api-access-qbp2v\") pod \"redhat-marketplace-wdw9k\" (UID: \"906c227f-a713-49d2-acfd-633646b6e0da\") " pod="openshift-marketplace/redhat-marketplace-wdw9k" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.598654 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906c227f-a713-49d2-acfd-633646b6e0da-utilities\") pod \"redhat-marketplace-wdw9k\" (UID: \"906c227f-a713-49d2-acfd-633646b6e0da\") " pod="openshift-marketplace/redhat-marketplace-wdw9k" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.598682 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906c227f-a713-49d2-acfd-633646b6e0da-catalog-content\") pod \"redhat-marketplace-wdw9k\" (UID: \"906c227f-a713-49d2-acfd-633646b6e0da\") " pod="openshift-marketplace/redhat-marketplace-wdw9k" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.599083 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906c227f-a713-49d2-acfd-633646b6e0da-catalog-content\") pod \"redhat-marketplace-wdw9k\" (UID: \"906c227f-a713-49d2-acfd-633646b6e0da\") " pod="openshift-marketplace/redhat-marketplace-wdw9k" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.599878 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-config-volume" (OuterVolumeSpecName: "config-volume") pod "ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3" (UID: "ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.600494 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906c227f-a713-49d2-acfd-633646b6e0da-utilities\") pod \"redhat-marketplace-wdw9k\" (UID: \"906c227f-a713-49d2-acfd-633646b6e0da\") " pod="openshift-marketplace/redhat-marketplace-wdw9k" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.605181 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3" (UID: "ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.605632 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-kube-api-access-g6b88" (OuterVolumeSpecName: "kube-api-access-g6b88") pod "ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3" (UID: "ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3"). InnerVolumeSpecName "kube-api-access-g6b88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.615696 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbp2v\" (UniqueName: \"kubernetes.io/projected/906c227f-a713-49d2-acfd-633646b6e0da-kube-api-access-qbp2v\") pod \"redhat-marketplace-wdw9k\" (UID: \"906c227f-a713-49d2-acfd-633646b6e0da\") " pod="openshift-marketplace/redhat-marketplace-wdw9k" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.681641 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdw9k" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.700098 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.700130 4706 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.700140 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6b88\" (UniqueName: \"kubernetes.io/projected/ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3-kube-api-access-g6b88\") on node \"crc\" DevicePath \"\"" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.761906 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sm9zj"] Dec 06 13:58:30 crc kubenswrapper[4706]: E1206 13:58:30.762218 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3" containerName="collect-profiles" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.762244 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3" containerName="collect-profiles" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.762659 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3" containerName="collect-profiles" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.764803 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sm9zj" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.769979 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sm9zj"] Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.903107 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcp94\" (UniqueName: \"kubernetes.io/projected/ebf6ef5b-30f8-4959-b07a-9b0512b19710-kube-api-access-hcp94\") pod \"redhat-marketplace-sm9zj\" (UID: \"ebf6ef5b-30f8-4959-b07a-9b0512b19710\") " pod="openshift-marketplace/redhat-marketplace-sm9zj" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.903677 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebf6ef5b-30f8-4959-b07a-9b0512b19710-utilities\") pod \"redhat-marketplace-sm9zj\" (UID: \"ebf6ef5b-30f8-4959-b07a-9b0512b19710\") " pod="openshift-marketplace/redhat-marketplace-sm9zj" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.903720 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebf6ef5b-30f8-4959-b07a-9b0512b19710-catalog-content\") pod \"redhat-marketplace-sm9zj\" (UID: \"ebf6ef5b-30f8-4959-b07a-9b0512b19710\") " pod="openshift-marketplace/redhat-marketplace-sm9zj" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.933462 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:30 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:30 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:30 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.933517 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:30 crc kubenswrapper[4706]: I1206 13:58:30.954601 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdw9k"] Dec 06 13:58:30 crc kubenswrapper[4706]: W1206 13:58:30.965766 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod906c227f_a713_49d2_acfd_633646b6e0da.slice/crio-7124ce5e15f7aa7359e7e501561a5f3f5982a75a865ea5ebe4dc3a29576f6a30 WatchSource:0}: Error finding container 7124ce5e15f7aa7359e7e501561a5f3f5982a75a865ea5ebe4dc3a29576f6a30: Status 404 returned error can't find the container with id 7124ce5e15f7aa7359e7e501561a5f3f5982a75a865ea5ebe4dc3a29576f6a30 Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.004667 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcp94\" (UniqueName: \"kubernetes.io/projected/ebf6ef5b-30f8-4959-b07a-9b0512b19710-kube-api-access-hcp94\") pod \"redhat-marketplace-sm9zj\" (UID: \"ebf6ef5b-30f8-4959-b07a-9b0512b19710\") " pod="openshift-marketplace/redhat-marketplace-sm9zj" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.004730 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebf6ef5b-30f8-4959-b07a-9b0512b19710-utilities\") pod \"redhat-marketplace-sm9zj\" (UID: \"ebf6ef5b-30f8-4959-b07a-9b0512b19710\") " pod="openshift-marketplace/redhat-marketplace-sm9zj" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.004755 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebf6ef5b-30f8-4959-b07a-9b0512b19710-catalog-content\") pod \"redhat-marketplace-sm9zj\" (UID: \"ebf6ef5b-30f8-4959-b07a-9b0512b19710\") " pod="openshift-marketplace/redhat-marketplace-sm9zj" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.005143 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebf6ef5b-30f8-4959-b07a-9b0512b19710-catalog-content\") pod \"redhat-marketplace-sm9zj\" (UID: \"ebf6ef5b-30f8-4959-b07a-9b0512b19710\") " pod="openshift-marketplace/redhat-marketplace-sm9zj" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.005319 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebf6ef5b-30f8-4959-b07a-9b0512b19710-utilities\") pod \"redhat-marketplace-sm9zj\" (UID: \"ebf6ef5b-30f8-4959-b07a-9b0512b19710\") " pod="openshift-marketplace/redhat-marketplace-sm9zj" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.033030 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcp94\" (UniqueName: \"kubernetes.io/projected/ebf6ef5b-30f8-4959-b07a-9b0512b19710-kube-api-access-hcp94\") pod \"redhat-marketplace-sm9zj\" (UID: \"ebf6ef5b-30f8-4959-b07a-9b0512b19710\") " pod="openshift-marketplace/redhat-marketplace-sm9zj" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.087112 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sm9zj" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.189520 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3904e406-8755-45fa-8088-93ecca405707","Type":"ContainerStarted","Data":"0ca600adb331e2abf7241a65f5f5f1c6c0dce20be5d93b8b624e1a1592da34f6"} Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.189562 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3904e406-8755-45fa-8088-93ecca405707","Type":"ContainerStarted","Data":"d87a0436101175c7a6d8e7dbdddcb1d1a876f3d7feb601668699c456b89eb1c6"} Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.195903 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdw9k" event={"ID":"906c227f-a713-49d2-acfd-633646b6e0da","Type":"ContainerStarted","Data":"7124ce5e15f7aa7359e7e501561a5f3f5982a75a865ea5ebe4dc3a29576f6a30"} Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.197451 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.197465 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417145-v956j" event={"ID":"ed3cbb3f-fda7-48cb-baa4-c6d1dc7257b3","Type":"ContainerDied","Data":"23e9db62230960184342a1cd4ca8f54fcbde0688a07f952b12632534cdf01351"} Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.197491 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23e9db62230960184342a1cd4ca8f54fcbde0688a07f952b12632534cdf01351" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.205183 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.205172051 podStartE2EDuration="2.205172051s" podCreationTimestamp="2025-12-06 13:58:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:58:31.202980595 +0000 UTC m=+154.038773452" watchObservedRunningTime="2025-12-06 13:58:31.205172051 +0000 UTC m=+154.040964908" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.558180 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x5pm6"] Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.560460 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5pm6" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.560508 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sm9zj"] Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.562789 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.562911 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x5pm6"] Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.623610 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-9jn7l" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.714241 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34cf831e-196b-46c2-a2a1-d985cdfd8861-catalog-content\") pod \"redhat-operators-x5pm6\" (UID: \"34cf831e-196b-46c2-a2a1-d985cdfd8861\") " pod="openshift-marketplace/redhat-operators-x5pm6" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.714333 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34cf831e-196b-46c2-a2a1-d985cdfd8861-utilities\") pod \"redhat-operators-x5pm6\" (UID: \"34cf831e-196b-46c2-a2a1-d985cdfd8861\") " pod="openshift-marketplace/redhat-operators-x5pm6" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.714662 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvbhh\" (UniqueName: \"kubernetes.io/projected/34cf831e-196b-46c2-a2a1-d985cdfd8861-kube-api-access-fvbhh\") pod \"redhat-operators-x5pm6\" (UID: \"34cf831e-196b-46c2-a2a1-d985cdfd8861\") " pod="openshift-marketplace/redhat-operators-x5pm6" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.815395 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvbhh\" (UniqueName: \"kubernetes.io/projected/34cf831e-196b-46c2-a2a1-d985cdfd8861-kube-api-access-fvbhh\") pod \"redhat-operators-x5pm6\" (UID: \"34cf831e-196b-46c2-a2a1-d985cdfd8861\") " pod="openshift-marketplace/redhat-operators-x5pm6" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.815450 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34cf831e-196b-46c2-a2a1-d985cdfd8861-catalog-content\") pod \"redhat-operators-x5pm6\" (UID: \"34cf831e-196b-46c2-a2a1-d985cdfd8861\") " pod="openshift-marketplace/redhat-operators-x5pm6" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.815484 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34cf831e-196b-46c2-a2a1-d985cdfd8861-utilities\") pod \"redhat-operators-x5pm6\" (UID: \"34cf831e-196b-46c2-a2a1-d985cdfd8861\") " pod="openshift-marketplace/redhat-operators-x5pm6" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.815937 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34cf831e-196b-46c2-a2a1-d985cdfd8861-utilities\") pod \"redhat-operators-x5pm6\" (UID: \"34cf831e-196b-46c2-a2a1-d985cdfd8861\") " pod="openshift-marketplace/redhat-operators-x5pm6" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.816177 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34cf831e-196b-46c2-a2a1-d985cdfd8861-catalog-content\") pod \"redhat-operators-x5pm6\" (UID: \"34cf831e-196b-46c2-a2a1-d985cdfd8861\") " pod="openshift-marketplace/redhat-operators-x5pm6" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.829115 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.829154 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.834482 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.840344 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvbhh\" (UniqueName: \"kubernetes.io/projected/34cf831e-196b-46c2-a2a1-d985cdfd8861-kube-api-access-fvbhh\") pod \"redhat-operators-x5pm6\" (UID: \"34cf831e-196b-46c2-a2a1-d985cdfd8861\") " pod="openshift-marketplace/redhat-operators-x5pm6" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.892990 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5pm6" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.929784 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.932065 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:31 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:31 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:31 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.932320 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.950656 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v9dc2"] Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.951616 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v9dc2" Dec 06 13:58:31 crc kubenswrapper[4706]: I1206 13:58:31.958187 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v9dc2"] Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.121184 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/437efad3-ab4d-4e73-86c1-79f182efabbb-utilities\") pod \"redhat-operators-v9dc2\" (UID: \"437efad3-ab4d-4e73-86c1-79f182efabbb\") " pod="openshift-marketplace/redhat-operators-v9dc2" Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.121221 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/437efad3-ab4d-4e73-86c1-79f182efabbb-catalog-content\") pod \"redhat-operators-v9dc2\" (UID: \"437efad3-ab4d-4e73-86c1-79f182efabbb\") " pod="openshift-marketplace/redhat-operators-v9dc2" Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.121285 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzsv2\" (UniqueName: \"kubernetes.io/projected/437efad3-ab4d-4e73-86c1-79f182efabbb-kube-api-access-nzsv2\") pod \"redhat-operators-v9dc2\" (UID: \"437efad3-ab4d-4e73-86c1-79f182efabbb\") " pod="openshift-marketplace/redhat-operators-v9dc2" Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.186392 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x5pm6"] Dec 06 13:58:32 crc kubenswrapper[4706]: W1206 13:58:32.201827 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34cf831e_196b_46c2_a2a1_d985cdfd8861.slice/crio-8aa6bb249443c49e315b16d56261c28ead42a575c350e18aa8a1b60072b9bbd4 WatchSource:0}: Error finding container 8aa6bb249443c49e315b16d56261c28ead42a575c350e18aa8a1b60072b9bbd4: Status 404 returned error can't find the container with id 8aa6bb249443c49e315b16d56261c28ead42a575c350e18aa8a1b60072b9bbd4 Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.208978 4706 generic.go:334] "Generic (PLEG): container finished" podID="ebf6ef5b-30f8-4959-b07a-9b0512b19710" containerID="cd7776a885ec55e84f412b74be8186a6444007a15547a7088274c1907e7c98b8" exitCode=0 Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.209170 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sm9zj" event={"ID":"ebf6ef5b-30f8-4959-b07a-9b0512b19710","Type":"ContainerDied","Data":"cd7776a885ec55e84f412b74be8186a6444007a15547a7088274c1907e7c98b8"} Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.209204 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sm9zj" event={"ID":"ebf6ef5b-30f8-4959-b07a-9b0512b19710","Type":"ContainerStarted","Data":"066a4b6604da17b9be68ab47f5d6a3804b1cab336abde94d43f5beb57f6355d7"} Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.214521 4706 generic.go:334] "Generic (PLEG): container finished" podID="3904e406-8755-45fa-8088-93ecca405707" containerID="0ca600adb331e2abf7241a65f5f5f1c6c0dce20be5d93b8b624e1a1592da34f6" exitCode=0 Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.214563 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3904e406-8755-45fa-8088-93ecca405707","Type":"ContainerDied","Data":"0ca600adb331e2abf7241a65f5f5f1c6c0dce20be5d93b8b624e1a1592da34f6"} Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.217249 4706 generic.go:334] "Generic (PLEG): container finished" podID="906c227f-a713-49d2-acfd-633646b6e0da" containerID="91195cfe1f00c83e86251aa226042c619b7f0b3ed291c91a38e02209edcc50ce" exitCode=0 Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.217427 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdw9k" event={"ID":"906c227f-a713-49d2-acfd-633646b6e0da","Type":"ContainerDied","Data":"91195cfe1f00c83e86251aa226042c619b7f0b3ed291c91a38e02209edcc50ce"} Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.221993 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzsv2\" (UniqueName: \"kubernetes.io/projected/437efad3-ab4d-4e73-86c1-79f182efabbb-kube-api-access-nzsv2\") pod \"redhat-operators-v9dc2\" (UID: \"437efad3-ab4d-4e73-86c1-79f182efabbb\") " pod="openshift-marketplace/redhat-operators-v9dc2" Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.222046 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/437efad3-ab4d-4e73-86c1-79f182efabbb-utilities\") pod \"redhat-operators-v9dc2\" (UID: \"437efad3-ab4d-4e73-86c1-79f182efabbb\") " pod="openshift-marketplace/redhat-operators-v9dc2" Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.222063 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/437efad3-ab4d-4e73-86c1-79f182efabbb-catalog-content\") pod \"redhat-operators-v9dc2\" (UID: \"437efad3-ab4d-4e73-86c1-79f182efabbb\") " pod="openshift-marketplace/redhat-operators-v9dc2" Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.222639 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/437efad3-ab4d-4e73-86c1-79f182efabbb-catalog-content\") pod \"redhat-operators-v9dc2\" (UID: \"437efad3-ab4d-4e73-86c1-79f182efabbb\") " pod="openshift-marketplace/redhat-operators-v9dc2" Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.223620 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/437efad3-ab4d-4e73-86c1-79f182efabbb-utilities\") pod \"redhat-operators-v9dc2\" (UID: \"437efad3-ab4d-4e73-86c1-79f182efabbb\") " pod="openshift-marketplace/redhat-operators-v9dc2" Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.223650 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-p8xpt" Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.243865 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzsv2\" (UniqueName: \"kubernetes.io/projected/437efad3-ab4d-4e73-86c1-79f182efabbb-kube-api-access-nzsv2\") pod \"redhat-operators-v9dc2\" (UID: \"437efad3-ab4d-4e73-86c1-79f182efabbb\") " pod="openshift-marketplace/redhat-operators-v9dc2" Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.256182 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.256215 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.258227 4706 patch_prober.go:28] interesting pod/console-f9d7485db-ldd62 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.258256 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-ldd62" podUID="532ce376-db1f-4770-a4cd-a30318242daf" containerName="console" probeResult="failure" output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.298431 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v9dc2" Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.655531 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v9dc2"] Dec 06 13:58:32 crc kubenswrapper[4706]: W1206 13:58:32.685350 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod437efad3_ab4d_4e73_86c1_79f182efabbb.slice/crio-a57dfaae493e5dbc17ecd71d8d9a55bd5e93a1c0e0a1a0491fe0a83826f9a1b9 WatchSource:0}: Error finding container a57dfaae493e5dbc17ecd71d8d9a55bd5e93a1c0e0a1a0491fe0a83826f9a1b9: Status 404 returned error can't find the container with id a57dfaae493e5dbc17ecd71d8d9a55bd5e93a1c0e0a1a0491fe0a83826f9a1b9 Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.933087 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:32 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:32 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:32 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:32 crc kubenswrapper[4706]: I1206 13:58:32.933376 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:33 crc kubenswrapper[4706]: I1206 13:58:33.240848 4706 generic.go:334] "Generic (PLEG): container finished" podID="34cf831e-196b-46c2-a2a1-d985cdfd8861" containerID="bbe5490c2e1844ce1d4095ad4ee7bcc4dc2548492ed6bc2933f96418e8667b5a" exitCode=0 Dec 06 13:58:33 crc kubenswrapper[4706]: I1206 13:58:33.240913 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5pm6" event={"ID":"34cf831e-196b-46c2-a2a1-d985cdfd8861","Type":"ContainerDied","Data":"bbe5490c2e1844ce1d4095ad4ee7bcc4dc2548492ed6bc2933f96418e8667b5a"} Dec 06 13:58:33 crc kubenswrapper[4706]: I1206 13:58:33.240939 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5pm6" event={"ID":"34cf831e-196b-46c2-a2a1-d985cdfd8861","Type":"ContainerStarted","Data":"8aa6bb249443c49e315b16d56261c28ead42a575c350e18aa8a1b60072b9bbd4"} Dec 06 13:58:33 crc kubenswrapper[4706]: I1206 13:58:33.246468 4706 generic.go:334] "Generic (PLEG): container finished" podID="437efad3-ab4d-4e73-86c1-79f182efabbb" containerID="ee7db2a27d3794da2faf8839a0d77e17852104bc4bbb231bcedf2e4772f91392" exitCode=0 Dec 06 13:58:33 crc kubenswrapper[4706]: I1206 13:58:33.247093 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v9dc2" event={"ID":"437efad3-ab4d-4e73-86c1-79f182efabbb","Type":"ContainerDied","Data":"ee7db2a27d3794da2faf8839a0d77e17852104bc4bbb231bcedf2e4772f91392"} Dec 06 13:58:33 crc kubenswrapper[4706]: I1206 13:58:33.247124 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v9dc2" event={"ID":"437efad3-ab4d-4e73-86c1-79f182efabbb","Type":"ContainerStarted","Data":"a57dfaae493e5dbc17ecd71d8d9a55bd5e93a1c0e0a1a0491fe0a83826f9a1b9"} Dec 06 13:58:33 crc kubenswrapper[4706]: I1206 13:58:33.640911 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 13:58:33 crc kubenswrapper[4706]: I1206 13:58:33.752054 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3904e406-8755-45fa-8088-93ecca405707-kubelet-dir\") pod \"3904e406-8755-45fa-8088-93ecca405707\" (UID: \"3904e406-8755-45fa-8088-93ecca405707\") " Dec 06 13:58:33 crc kubenswrapper[4706]: I1206 13:58:33.752128 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3904e406-8755-45fa-8088-93ecca405707-kube-api-access\") pod \"3904e406-8755-45fa-8088-93ecca405707\" (UID: \"3904e406-8755-45fa-8088-93ecca405707\") " Dec 06 13:58:33 crc kubenswrapper[4706]: I1206 13:58:33.753670 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3904e406-8755-45fa-8088-93ecca405707-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3904e406-8755-45fa-8088-93ecca405707" (UID: "3904e406-8755-45fa-8088-93ecca405707"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 13:58:33 crc kubenswrapper[4706]: I1206 13:58:33.758227 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3904e406-8755-45fa-8088-93ecca405707-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3904e406-8755-45fa-8088-93ecca405707" (UID: "3904e406-8755-45fa-8088-93ecca405707"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:58:33 crc kubenswrapper[4706]: I1206 13:58:33.853421 4706 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3904e406-8755-45fa-8088-93ecca405707-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 13:58:33 crc kubenswrapper[4706]: I1206 13:58:33.853448 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3904e406-8755-45fa-8088-93ecca405707-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 13:58:33 crc kubenswrapper[4706]: I1206 13:58:33.935457 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:33 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:33 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:33 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:33 crc kubenswrapper[4706]: I1206 13:58:33.935686 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:34 crc kubenswrapper[4706]: I1206 13:58:34.287665 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3904e406-8755-45fa-8088-93ecca405707","Type":"ContainerDied","Data":"d87a0436101175c7a6d8e7dbdddcb1d1a876f3d7feb601668699c456b89eb1c6"} Dec 06 13:58:34 crc kubenswrapper[4706]: I1206 13:58:34.287705 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d87a0436101175c7a6d8e7dbdddcb1d1a876f3d7feb601668699c456b89eb1c6" Dec 06 13:58:34 crc kubenswrapper[4706]: I1206 13:58:34.287774 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 13:58:34 crc kubenswrapper[4706]: I1206 13:58:34.919643 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 13:58:34 crc kubenswrapper[4706]: E1206 13:58:34.919841 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3904e406-8755-45fa-8088-93ecca405707" containerName="pruner" Dec 06 13:58:34 crc kubenswrapper[4706]: I1206 13:58:34.919851 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3904e406-8755-45fa-8088-93ecca405707" containerName="pruner" Dec 06 13:58:34 crc kubenswrapper[4706]: I1206 13:58:34.919941 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3904e406-8755-45fa-8088-93ecca405707" containerName="pruner" Dec 06 13:58:34 crc kubenswrapper[4706]: I1206 13:58:34.920258 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 13:58:34 crc kubenswrapper[4706]: I1206 13:58:34.925211 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 13:58:34 crc kubenswrapper[4706]: I1206 13:58:34.925910 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 13:58:34 crc kubenswrapper[4706]: I1206 13:58:34.928864 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 13:58:34 crc kubenswrapper[4706]: I1206 13:58:34.932669 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:34 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:34 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:34 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:34 crc kubenswrapper[4706]: I1206 13:58:34.933234 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:35 crc kubenswrapper[4706]: I1206 13:58:35.092851 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7825cac6-f0a3-4589-be48-c6ced1f84860-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7825cac6-f0a3-4589-be48-c6ced1f84860\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 13:58:35 crc kubenswrapper[4706]: I1206 13:58:35.092975 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7825cac6-f0a3-4589-be48-c6ced1f84860-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7825cac6-f0a3-4589-be48-c6ced1f84860\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 13:58:35 crc kubenswrapper[4706]: I1206 13:58:35.193755 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7825cac6-f0a3-4589-be48-c6ced1f84860-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7825cac6-f0a3-4589-be48-c6ced1f84860\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 13:58:35 crc kubenswrapper[4706]: I1206 13:58:35.193833 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7825cac6-f0a3-4589-be48-c6ced1f84860-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7825cac6-f0a3-4589-be48-c6ced1f84860\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 13:58:35 crc kubenswrapper[4706]: I1206 13:58:35.193901 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7825cac6-f0a3-4589-be48-c6ced1f84860-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7825cac6-f0a3-4589-be48-c6ced1f84860\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 13:58:35 crc kubenswrapper[4706]: I1206 13:58:35.218270 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7825cac6-f0a3-4589-be48-c6ced1f84860-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7825cac6-f0a3-4589-be48-c6ced1f84860\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 13:58:35 crc kubenswrapper[4706]: I1206 13:58:35.240919 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 13:58:35 crc kubenswrapper[4706]: I1206 13:58:35.433403 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 13:58:35 crc kubenswrapper[4706]: I1206 13:58:35.466817 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 13:58:35 crc kubenswrapper[4706]: W1206 13:58:35.478667 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7825cac6_f0a3_4589_be48_c6ced1f84860.slice/crio-ec12b0d07530bd47f4014b469ec188b61b33bc7e09a422a76cb62c174f8da732 WatchSource:0}: Error finding container ec12b0d07530bd47f4014b469ec188b61b33bc7e09a422a76cb62c174f8da732: Status 404 returned error can't find the container with id ec12b0d07530bd47f4014b469ec188b61b33bc7e09a422a76cb62c174f8da732 Dec 06 13:58:35 crc kubenswrapper[4706]: I1206 13:58:35.939944 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:35 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:35 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:35 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:35 crc kubenswrapper[4706]: I1206 13:58:35.940285 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:36 crc kubenswrapper[4706]: I1206 13:58:36.302408 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7825cac6-f0a3-4589-be48-c6ced1f84860","Type":"ContainerStarted","Data":"c3b7464a6425ee846dcf1488da9749bce51348a94d2cb9da65a4187714f09b81"} Dec 06 13:58:36 crc kubenswrapper[4706]: I1206 13:58:36.302451 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7825cac6-f0a3-4589-be48-c6ced1f84860","Type":"ContainerStarted","Data":"ec12b0d07530bd47f4014b469ec188b61b33bc7e09a422a76cb62c174f8da732"} Dec 06 13:58:36 crc kubenswrapper[4706]: I1206 13:58:36.937424 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:36 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:36 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:36 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:36 crc kubenswrapper[4706]: I1206 13:58:36.937717 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:37 crc kubenswrapper[4706]: I1206 13:58:37.317387 4706 generic.go:334] "Generic (PLEG): container finished" podID="7825cac6-f0a3-4589-be48-c6ced1f84860" containerID="c3b7464a6425ee846dcf1488da9749bce51348a94d2cb9da65a4187714f09b81" exitCode=0 Dec 06 13:58:37 crc kubenswrapper[4706]: I1206 13:58:37.317440 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7825cac6-f0a3-4589-be48-c6ced1f84860","Type":"ContainerDied","Data":"c3b7464a6425ee846dcf1488da9749bce51348a94d2cb9da65a4187714f09b81"} Dec 06 13:58:37 crc kubenswrapper[4706]: I1206 13:58:37.774360 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-m9fbd" Dec 06 13:58:37 crc kubenswrapper[4706]: I1206 13:58:37.932821 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:37 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:37 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:37 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:37 crc kubenswrapper[4706]: I1206 13:58:37.932882 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:38 crc kubenswrapper[4706]: I1206 13:58:38.932390 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:38 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:38 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:38 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:38 crc kubenswrapper[4706]: I1206 13:58:38.932799 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:39 crc kubenswrapper[4706]: I1206 13:58:39.265657 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs\") pod \"network-metrics-daemon-hhkhk\" (UID: \"5443fc82-2d54-43df-88da-043c2eb77238\") " pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:58:39 crc kubenswrapper[4706]: I1206 13:58:39.279380 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5443fc82-2d54-43df-88da-043c2eb77238-metrics-certs\") pod \"network-metrics-daemon-hhkhk\" (UID: \"5443fc82-2d54-43df-88da-043c2eb77238\") " pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:58:39 crc kubenswrapper[4706]: I1206 13:58:39.436397 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhkhk" Dec 06 13:58:39 crc kubenswrapper[4706]: I1206 13:58:39.931956 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:39 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:39 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:39 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:39 crc kubenswrapper[4706]: I1206 13:58:39.932023 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:40 crc kubenswrapper[4706]: I1206 13:58:40.931395 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:40 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:40 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:40 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:40 crc kubenswrapper[4706]: I1206 13:58:40.931447 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:41 crc kubenswrapper[4706]: I1206 13:58:41.934189 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:41 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:41 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:41 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:41 crc kubenswrapper[4706]: I1206 13:58:41.934453 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:42 crc kubenswrapper[4706]: I1206 13:58:42.256383 4706 patch_prober.go:28] interesting pod/console-f9d7485db-ldd62 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 06 13:58:42 crc kubenswrapper[4706]: I1206 13:58:42.256441 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-ldd62" podUID="532ce376-db1f-4770-a4cd-a30318242daf" containerName="console" probeResult="failure" output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 06 13:58:42 crc kubenswrapper[4706]: I1206 13:58:42.931790 4706 patch_prober.go:28] interesting pod/router-default-5444994796-r94jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 13:58:42 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Dec 06 13:58:42 crc kubenswrapper[4706]: [+]process-running ok Dec 06 13:58:42 crc kubenswrapper[4706]: healthz check failed Dec 06 13:58:42 crc kubenswrapper[4706]: I1206 13:58:42.932099 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r94jx" podUID="b9f16df2-5700-43c2-9488-78c3a9d8bcd6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 13:58:43 crc kubenswrapper[4706]: I1206 13:58:43.874828 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 13:58:43 crc kubenswrapper[4706]: I1206 13:58:43.933274 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:43 crc kubenswrapper[4706]: I1206 13:58:43.937375 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-r94jx" Dec 06 13:58:44 crc kubenswrapper[4706]: I1206 13:58:44.031730 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7825cac6-f0a3-4589-be48-c6ced1f84860-kube-api-access\") pod \"7825cac6-f0a3-4589-be48-c6ced1f84860\" (UID: \"7825cac6-f0a3-4589-be48-c6ced1f84860\") " Dec 06 13:58:44 crc kubenswrapper[4706]: I1206 13:58:44.031825 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7825cac6-f0a3-4589-be48-c6ced1f84860-kubelet-dir\") pod \"7825cac6-f0a3-4589-be48-c6ced1f84860\" (UID: \"7825cac6-f0a3-4589-be48-c6ced1f84860\") " Dec 06 13:58:44 crc kubenswrapper[4706]: I1206 13:58:44.032113 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7825cac6-f0a3-4589-be48-c6ced1f84860-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7825cac6-f0a3-4589-be48-c6ced1f84860" (UID: "7825cac6-f0a3-4589-be48-c6ced1f84860"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 13:58:44 crc kubenswrapper[4706]: I1206 13:58:44.032345 4706 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7825cac6-f0a3-4589-be48-c6ced1f84860-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 13:58:44 crc kubenswrapper[4706]: I1206 13:58:44.038571 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7825cac6-f0a3-4589-be48-c6ced1f84860-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7825cac6-f0a3-4589-be48-c6ced1f84860" (UID: "7825cac6-f0a3-4589-be48-c6ced1f84860"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:58:44 crc kubenswrapper[4706]: I1206 13:58:44.133426 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7825cac6-f0a3-4589-be48-c6ced1f84860-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 13:58:44 crc kubenswrapper[4706]: I1206 13:58:44.381999 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 13:58:44 crc kubenswrapper[4706]: I1206 13:58:44.382017 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7825cac6-f0a3-4589-be48-c6ced1f84860","Type":"ContainerDied","Data":"ec12b0d07530bd47f4014b469ec188b61b33bc7e09a422a76cb62c174f8da732"} Dec 06 13:58:44 crc kubenswrapper[4706]: I1206 13:58:44.382070 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec12b0d07530bd47f4014b469ec188b61b33bc7e09a422a76cb62c174f8da732" Dec 06 13:58:44 crc kubenswrapper[4706]: I1206 13:58:44.579906 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 13:58:44 crc kubenswrapper[4706]: I1206 13:58:44.579954 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 13:58:49 crc kubenswrapper[4706]: I1206 13:58:49.173027 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 13:58:52 crc kubenswrapper[4706]: I1206 13:58:52.259298 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:52 crc kubenswrapper[4706]: I1206 13:58:52.264691 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-ldd62" Dec 06 13:58:59 crc kubenswrapper[4706]: E1206 13:58:59.057272 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 13:58:59 crc kubenswrapper[4706]: E1206 13:58:59.058680 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2l68w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-gx45h_openshift-marketplace(06e9400a-9432-4f75-af53-78b3b3facd7a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 13:58:59 crc kubenswrapper[4706]: E1206 13:58:59.059948 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-gx45h" podUID="06e9400a-9432-4f75-af53-78b3b3facd7a" Dec 06 13:59:02 crc kubenswrapper[4706]: I1206 13:59:02.662587 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dp75h" Dec 06 13:59:04 crc kubenswrapper[4706]: E1206 13:59:04.544547 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-gx45h" podUID="06e9400a-9432-4f75-af53-78b3b3facd7a" Dec 06 13:59:04 crc kubenswrapper[4706]: I1206 13:59:04.972786 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 13:59:06 crc kubenswrapper[4706]: E1206 13:59:06.970191 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 13:59:06 crc kubenswrapper[4706]: E1206 13:59:06.970654 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6p4zf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-kslt9_openshift-marketplace(1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 13:59:06 crc kubenswrapper[4706]: E1206 13:59:06.972055 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-kslt9" podUID="1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" Dec 06 13:59:06 crc kubenswrapper[4706]: E1206 13:59:06.982012 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 06 13:59:06 crc kubenswrapper[4706]: E1206 13:59:06.982177 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qbp2v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-wdw9k_openshift-marketplace(906c227f-a713-49d2-acfd-633646b6e0da): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 13:59:06 crc kubenswrapper[4706]: E1206 13:59:06.983609 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-wdw9k" podUID="906c227f-a713-49d2-acfd-633646b6e0da" Dec 06 13:59:09 crc kubenswrapper[4706]: I1206 13:59:09.721009 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 13:59:09 crc kubenswrapper[4706]: E1206 13:59:09.721468 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7825cac6-f0a3-4589-be48-c6ced1f84860" containerName="pruner" Dec 06 13:59:09 crc kubenswrapper[4706]: I1206 13:59:09.721478 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="7825cac6-f0a3-4589-be48-c6ced1f84860" containerName="pruner" Dec 06 13:59:09 crc kubenswrapper[4706]: I1206 13:59:09.721567 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="7825cac6-f0a3-4589-be48-c6ced1f84860" containerName="pruner" Dec 06 13:59:09 crc kubenswrapper[4706]: I1206 13:59:09.725717 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 13:59:09 crc kubenswrapper[4706]: I1206 13:59:09.728245 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 13:59:09 crc kubenswrapper[4706]: I1206 13:59:09.728452 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 13:59:09 crc kubenswrapper[4706]: I1206 13:59:09.730399 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 13:59:09 crc kubenswrapper[4706]: I1206 13:59:09.817036 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8ceeceb1-f929-4022-9c4c-68d608dfdaaa-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8ceeceb1-f929-4022-9c4c-68d608dfdaaa\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 13:59:09 crc kubenswrapper[4706]: I1206 13:59:09.817088 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8ceeceb1-f929-4022-9c4c-68d608dfdaaa-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8ceeceb1-f929-4022-9c4c-68d608dfdaaa\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 13:59:09 crc kubenswrapper[4706]: I1206 13:59:09.918362 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8ceeceb1-f929-4022-9c4c-68d608dfdaaa-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8ceeceb1-f929-4022-9c4c-68d608dfdaaa\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 13:59:09 crc kubenswrapper[4706]: I1206 13:59:09.918693 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8ceeceb1-f929-4022-9c4c-68d608dfdaaa-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8ceeceb1-f929-4022-9c4c-68d608dfdaaa\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 13:59:09 crc kubenswrapper[4706]: I1206 13:59:09.918604 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8ceeceb1-f929-4022-9c4c-68d608dfdaaa-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8ceeceb1-f929-4022-9c4c-68d608dfdaaa\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 13:59:09 crc kubenswrapper[4706]: I1206 13:59:09.938230 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8ceeceb1-f929-4022-9c4c-68d608dfdaaa-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8ceeceb1-f929-4022-9c4c-68d608dfdaaa\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 13:59:09 crc kubenswrapper[4706]: E1206 13:59:09.950703 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdw9k" podUID="906c227f-a713-49d2-acfd-633646b6e0da" Dec 06 13:59:09 crc kubenswrapper[4706]: E1206 13:59:09.950718 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-kslt9" podUID="1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" Dec 06 13:59:10 crc kubenswrapper[4706]: E1206 13:59:10.020443 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 06 13:59:10 crc kubenswrapper[4706]: E1206 13:59:10.020907 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fvbhh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-x5pm6_openshift-marketplace(34cf831e-196b-46c2-a2a1-d985cdfd8861): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 13:59:10 crc kubenswrapper[4706]: E1206 13:59:10.022109 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-x5pm6" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" Dec 06 13:59:10 crc kubenswrapper[4706]: E1206 13:59:10.036027 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 06 13:59:10 crc kubenswrapper[4706]: E1206 13:59:10.036183 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hcp94,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-sm9zj_openshift-marketplace(ebf6ef5b-30f8-4959-b07a-9b0512b19710): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 13:59:10 crc kubenswrapper[4706]: E1206 13:59:10.037513 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-sm9zj" podUID="ebf6ef5b-30f8-4959-b07a-9b0512b19710" Dec 06 13:59:10 crc kubenswrapper[4706]: I1206 13:59:10.082845 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 13:59:11 crc kubenswrapper[4706]: E1206 13:59:11.648122 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-sm9zj" podUID="ebf6ef5b-30f8-4959-b07a-9b0512b19710" Dec 06 13:59:11 crc kubenswrapper[4706]: E1206 13:59:11.648271 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x5pm6" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" Dec 06 13:59:11 crc kubenswrapper[4706]: E1206 13:59:11.725388 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 06 13:59:11 crc kubenswrapper[4706]: E1206 13:59:11.725563 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9w4bj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zhlvc_openshift-marketplace(d12f4107-0d89-4d3c-958f-7355fbd27575): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 13:59:11 crc kubenswrapper[4706]: E1206 13:59:11.727219 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zhlvc" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" Dec 06 13:59:11 crc kubenswrapper[4706]: E1206 13:59:11.739370 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 06 13:59:11 crc kubenswrapper[4706]: E1206 13:59:11.739569 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nzsv2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-v9dc2_openshift-marketplace(437efad3-ab4d-4e73-86c1-79f182efabbb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 13:59:11 crc kubenswrapper[4706]: E1206 13:59:11.741007 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-v9dc2" podUID="437efad3-ab4d-4e73-86c1-79f182efabbb" Dec 06 13:59:11 crc kubenswrapper[4706]: E1206 13:59:11.766630 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 06 13:59:11 crc kubenswrapper[4706]: E1206 13:59:11.767328 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fxrzw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-hjkcq_openshift-marketplace(93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 13:59:11 crc kubenswrapper[4706]: E1206 13:59:11.770253 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-hjkcq" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" Dec 06 13:59:12 crc kubenswrapper[4706]: I1206 13:59:12.114487 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 13:59:12 crc kubenswrapper[4706]: W1206 13:59:12.122782 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod8ceeceb1_f929_4022_9c4c_68d608dfdaaa.slice/crio-d2d81eb42991b521c98f1d9fda09be642f0a94b0d74630bcc1ba64fc9ef16cae WatchSource:0}: Error finding container d2d81eb42991b521c98f1d9fda09be642f0a94b0d74630bcc1ba64fc9ef16cae: Status 404 returned error can't find the container with id d2d81eb42991b521c98f1d9fda09be642f0a94b0d74630bcc1ba64fc9ef16cae Dec 06 13:59:12 crc kubenswrapper[4706]: I1206 13:59:12.140672 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hhkhk"] Dec 06 13:59:12 crc kubenswrapper[4706]: W1206 13:59:12.149855 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5443fc82_2d54_43df_88da_043c2eb77238.slice/crio-61e00abcc36805a8e27bd20c0a0abb553722a70b2f706c44164c0b8e15c3d0a9 WatchSource:0}: Error finding container 61e00abcc36805a8e27bd20c0a0abb553722a70b2f706c44164c0b8e15c3d0a9: Status 404 returned error can't find the container with id 61e00abcc36805a8e27bd20c0a0abb553722a70b2f706c44164c0b8e15c3d0a9 Dec 06 13:59:12 crc kubenswrapper[4706]: I1206 13:59:12.599263 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"8ceeceb1-f929-4022-9c4c-68d608dfdaaa","Type":"ContainerStarted","Data":"002a5812bb974233d8f98a75b3180bf265eb68c2d71ee281c1860cd30991186d"} Dec 06 13:59:12 crc kubenswrapper[4706]: I1206 13:59:12.599651 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"8ceeceb1-f929-4022-9c4c-68d608dfdaaa","Type":"ContainerStarted","Data":"d2d81eb42991b521c98f1d9fda09be642f0a94b0d74630bcc1ba64fc9ef16cae"} Dec 06 13:59:12 crc kubenswrapper[4706]: I1206 13:59:12.600974 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" event={"ID":"5443fc82-2d54-43df-88da-043c2eb77238","Type":"ContainerStarted","Data":"250f0131666eedb094ff622a3cc3b626cf76f680a80b8f833901014b0b9f727c"} Dec 06 13:59:12 crc kubenswrapper[4706]: I1206 13:59:12.601086 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" event={"ID":"5443fc82-2d54-43df-88da-043c2eb77238","Type":"ContainerStarted","Data":"61e00abcc36805a8e27bd20c0a0abb553722a70b2f706c44164c0b8e15c3d0a9"} Dec 06 13:59:12 crc kubenswrapper[4706]: E1206 13:59:12.602697 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-v9dc2" podUID="437efad3-ab4d-4e73-86c1-79f182efabbb" Dec 06 13:59:12 crc kubenswrapper[4706]: E1206 13:59:12.602596 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-hjkcq" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" Dec 06 13:59:12 crc kubenswrapper[4706]: E1206 13:59:12.610050 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zhlvc" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" Dec 06 13:59:12 crc kubenswrapper[4706]: I1206 13:59:12.616239 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=3.616193632 podStartE2EDuration="3.616193632s" podCreationTimestamp="2025-12-06 13:59:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:59:12.610295939 +0000 UTC m=+195.446088796" watchObservedRunningTime="2025-12-06 13:59:12.616193632 +0000 UTC m=+195.451986499" Dec 06 13:59:13 crc kubenswrapper[4706]: I1206 13:59:13.626111 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hhkhk" event={"ID":"5443fc82-2d54-43df-88da-043c2eb77238","Type":"ContainerStarted","Data":"06df14b6c8261c374aa5c2e031cbb784f1a2dbe320b14180689d7231f15065bc"} Dec 06 13:59:13 crc kubenswrapper[4706]: I1206 13:59:13.629657 4706 generic.go:334] "Generic (PLEG): container finished" podID="8ceeceb1-f929-4022-9c4c-68d608dfdaaa" containerID="002a5812bb974233d8f98a75b3180bf265eb68c2d71ee281c1860cd30991186d" exitCode=0 Dec 06 13:59:13 crc kubenswrapper[4706]: I1206 13:59:13.629720 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"8ceeceb1-f929-4022-9c4c-68d608dfdaaa","Type":"ContainerDied","Data":"002a5812bb974233d8f98a75b3180bf265eb68c2d71ee281c1860cd30991186d"} Dec 06 13:59:13 crc kubenswrapper[4706]: I1206 13:59:13.650720 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-hhkhk" podStartSLOduration=177.650700763 podStartE2EDuration="2m57.650700763s" podCreationTimestamp="2025-12-06 13:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:59:13.648357403 +0000 UTC m=+196.484150260" watchObservedRunningTime="2025-12-06 13:59:13.650700763 +0000 UTC m=+196.486493620" Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.523849 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.524558 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.530614 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.579871 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.580239 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.685366 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/40d27479-a513-43d4-8ad7-9a4daa8a29d7-kubelet-dir\") pod \"installer-9-crc\" (UID: \"40d27479-a513-43d4-8ad7-9a4daa8a29d7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.685424 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/40d27479-a513-43d4-8ad7-9a4daa8a29d7-var-lock\") pod \"installer-9-crc\" (UID: \"40d27479-a513-43d4-8ad7-9a4daa8a29d7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.685478 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/40d27479-a513-43d4-8ad7-9a4daa8a29d7-kube-api-access\") pod \"installer-9-crc\" (UID: \"40d27479-a513-43d4-8ad7-9a4daa8a29d7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.787333 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/40d27479-a513-43d4-8ad7-9a4daa8a29d7-kubelet-dir\") pod \"installer-9-crc\" (UID: \"40d27479-a513-43d4-8ad7-9a4daa8a29d7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.787454 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/40d27479-a513-43d4-8ad7-9a4daa8a29d7-var-lock\") pod \"installer-9-crc\" (UID: \"40d27479-a513-43d4-8ad7-9a4daa8a29d7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.787535 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/40d27479-a513-43d4-8ad7-9a4daa8a29d7-kube-api-access\") pod \"installer-9-crc\" (UID: \"40d27479-a513-43d4-8ad7-9a4daa8a29d7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.788059 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/40d27479-a513-43d4-8ad7-9a4daa8a29d7-var-lock\") pod \"installer-9-crc\" (UID: \"40d27479-a513-43d4-8ad7-9a4daa8a29d7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.788076 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/40d27479-a513-43d4-8ad7-9a4daa8a29d7-kubelet-dir\") pod \"installer-9-crc\" (UID: \"40d27479-a513-43d4-8ad7-9a4daa8a29d7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.812594 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/40d27479-a513-43d4-8ad7-9a4daa8a29d7-kube-api-access\") pod \"installer-9-crc\" (UID: \"40d27479-a513-43d4-8ad7-9a4daa8a29d7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.888528 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.902426 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.989941 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8ceeceb1-f929-4022-9c4c-68d608dfdaaa-kubelet-dir\") pod \"8ceeceb1-f929-4022-9c4c-68d608dfdaaa\" (UID: \"8ceeceb1-f929-4022-9c4c-68d608dfdaaa\") " Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.990046 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8ceeceb1-f929-4022-9c4c-68d608dfdaaa-kube-api-access\") pod \"8ceeceb1-f929-4022-9c4c-68d608dfdaaa\" (UID: \"8ceeceb1-f929-4022-9c4c-68d608dfdaaa\") " Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.990078 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ceeceb1-f929-4022-9c4c-68d608dfdaaa-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "8ceeceb1-f929-4022-9c4c-68d608dfdaaa" (UID: "8ceeceb1-f929-4022-9c4c-68d608dfdaaa"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.990241 4706 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8ceeceb1-f929-4022-9c4c-68d608dfdaaa-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:14 crc kubenswrapper[4706]: I1206 13:59:14.992724 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ceeceb1-f929-4022-9c4c-68d608dfdaaa-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "8ceeceb1-f929-4022-9c4c-68d608dfdaaa" (UID: "8ceeceb1-f929-4022-9c4c-68d608dfdaaa"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:59:15 crc kubenswrapper[4706]: I1206 13:59:15.103142 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8ceeceb1-f929-4022-9c4c-68d608dfdaaa-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:15 crc kubenswrapper[4706]: I1206 13:59:15.196707 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l55q6"] Dec 06 13:59:15 crc kubenswrapper[4706]: I1206 13:59:15.331435 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 13:59:15 crc kubenswrapper[4706]: W1206 13:59:15.344319 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod40d27479_a513_43d4_8ad7_9a4daa8a29d7.slice/crio-f2f91e9bce6f725c95b4835dc183f3f7e0c5179be7b62ed525d943a415873ba3 WatchSource:0}: Error finding container f2f91e9bce6f725c95b4835dc183f3f7e0c5179be7b62ed525d943a415873ba3: Status 404 returned error can't find the container with id f2f91e9bce6f725c95b4835dc183f3f7e0c5179be7b62ed525d943a415873ba3 Dec 06 13:59:15 crc kubenswrapper[4706]: I1206 13:59:15.641118 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"40d27479-a513-43d4-8ad7-9a4daa8a29d7","Type":"ContainerStarted","Data":"f2f91e9bce6f725c95b4835dc183f3f7e0c5179be7b62ed525d943a415873ba3"} Dec 06 13:59:15 crc kubenswrapper[4706]: I1206 13:59:15.642476 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"8ceeceb1-f929-4022-9c4c-68d608dfdaaa","Type":"ContainerDied","Data":"d2d81eb42991b521c98f1d9fda09be642f0a94b0d74630bcc1ba64fc9ef16cae"} Dec 06 13:59:15 crc kubenswrapper[4706]: I1206 13:59:15.642513 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2d81eb42991b521c98f1d9fda09be642f0a94b0d74630bcc1ba64fc9ef16cae" Dec 06 13:59:15 crc kubenswrapper[4706]: I1206 13:59:15.642523 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 13:59:16 crc kubenswrapper[4706]: I1206 13:59:16.651951 4706 generic.go:334] "Generic (PLEG): container finished" podID="06e9400a-9432-4f75-af53-78b3b3facd7a" containerID="6dfefcbf013ff43a6bfcc38e1ba8427c10e0249435ade86f54bd4c27b8c88b14" exitCode=0 Dec 06 13:59:16 crc kubenswrapper[4706]: I1206 13:59:16.652018 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gx45h" event={"ID":"06e9400a-9432-4f75-af53-78b3b3facd7a","Type":"ContainerDied","Data":"6dfefcbf013ff43a6bfcc38e1ba8427c10e0249435ade86f54bd4c27b8c88b14"} Dec 06 13:59:16 crc kubenswrapper[4706]: I1206 13:59:16.656446 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"40d27479-a513-43d4-8ad7-9a4daa8a29d7","Type":"ContainerStarted","Data":"c671b8c8691a77d80e9cd9d6fcce6bef62c242089e507629952c27a68739df84"} Dec 06 13:59:17 crc kubenswrapper[4706]: I1206 13:59:17.662523 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gx45h" event={"ID":"06e9400a-9432-4f75-af53-78b3b3facd7a","Type":"ContainerStarted","Data":"8285d11499d6bbaad4a4a9e05fcb3e203134e82112e069839dac6a72c5be5e16"} Dec 06 13:59:17 crc kubenswrapper[4706]: I1206 13:59:17.698362 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gx45h" podStartSLOduration=2.847118327 podStartE2EDuration="49.698340026s" podCreationTimestamp="2025-12-06 13:58:28 +0000 UTC" firstStartedPulling="2025-12-06 13:58:30.173529473 +0000 UTC m=+153.009322370" lastFinishedPulling="2025-12-06 13:59:17.024751202 +0000 UTC m=+199.860544069" observedRunningTime="2025-12-06 13:59:17.694486603 +0000 UTC m=+200.530279490" watchObservedRunningTime="2025-12-06 13:59:17.698340026 +0000 UTC m=+200.534132923" Dec 06 13:59:17 crc kubenswrapper[4706]: I1206 13:59:17.706159 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.7061378339999997 podStartE2EDuration="3.706137834s" podCreationTimestamp="2025-12-06 13:59:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:59:16.682658493 +0000 UTC m=+199.518451350" watchObservedRunningTime="2025-12-06 13:59:17.706137834 +0000 UTC m=+200.541930731" Dec 06 13:59:19 crc kubenswrapper[4706]: I1206 13:59:19.304706 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gx45h" Dec 06 13:59:19 crc kubenswrapper[4706]: I1206 13:59:19.305013 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gx45h" Dec 06 13:59:19 crc kubenswrapper[4706]: I1206 13:59:19.370764 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gx45h" Dec 06 13:59:23 crc kubenswrapper[4706]: I1206 13:59:23.700648 4706 generic.go:334] "Generic (PLEG): container finished" podID="ebf6ef5b-30f8-4959-b07a-9b0512b19710" containerID="1a8b649e0b36041830affaf539efee5bb01292cf814baecd3da267e883f0326b" exitCode=0 Dec 06 13:59:23 crc kubenswrapper[4706]: I1206 13:59:23.700760 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sm9zj" event={"ID":"ebf6ef5b-30f8-4959-b07a-9b0512b19710","Type":"ContainerDied","Data":"1a8b649e0b36041830affaf539efee5bb01292cf814baecd3da267e883f0326b"} Dec 06 13:59:24 crc kubenswrapper[4706]: I1206 13:59:24.710424 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sm9zj" event={"ID":"ebf6ef5b-30f8-4959-b07a-9b0512b19710","Type":"ContainerStarted","Data":"8a3c783d310d954db81fa89ba90c77cb5f46dfc7b15ecc13a4425c31c1f841d9"} Dec 06 13:59:24 crc kubenswrapper[4706]: I1206 13:59:24.712690 4706 generic.go:334] "Generic (PLEG): container finished" podID="1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" containerID="8571919fedef679e1130f6f1ddbad192d56d022c63105ccf1a58e8843a77353e" exitCode=0 Dec 06 13:59:24 crc kubenswrapper[4706]: I1206 13:59:24.712766 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kslt9" event={"ID":"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2","Type":"ContainerDied","Data":"8571919fedef679e1130f6f1ddbad192d56d022c63105ccf1a58e8843a77353e"} Dec 06 13:59:24 crc kubenswrapper[4706]: I1206 13:59:24.714277 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v9dc2" event={"ID":"437efad3-ab4d-4e73-86c1-79f182efabbb","Type":"ContainerStarted","Data":"0cc5677112f9507ab5a0ad6e8958e05b3eb0639161748ef868b8cf1178a58fad"} Dec 06 13:59:24 crc kubenswrapper[4706]: I1206 13:59:24.751239 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sm9zj" podStartSLOduration=2.845191773 podStartE2EDuration="54.751223462s" podCreationTimestamp="2025-12-06 13:58:30 +0000 UTC" firstStartedPulling="2025-12-06 13:58:32.213466076 +0000 UTC m=+155.049258923" lastFinishedPulling="2025-12-06 13:59:24.119497755 +0000 UTC m=+206.955290612" observedRunningTime="2025-12-06 13:59:24.728864505 +0000 UTC m=+207.564657352" watchObservedRunningTime="2025-12-06 13:59:24.751223462 +0000 UTC m=+207.587016319" Dec 06 13:59:25 crc kubenswrapper[4706]: I1206 13:59:25.721354 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kslt9" event={"ID":"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2","Type":"ContainerStarted","Data":"87ffbb89d97e84a1396792279ee8676f1771b30d9cdafddae97a066f3965e14d"} Dec 06 13:59:25 crc kubenswrapper[4706]: I1206 13:59:25.724097 4706 generic.go:334] "Generic (PLEG): container finished" podID="906c227f-a713-49d2-acfd-633646b6e0da" containerID="19d38017e092899a7cfd0aa5a92e4bd96a09ce9a9759dc303fb55cbd529fceba" exitCode=0 Dec 06 13:59:25 crc kubenswrapper[4706]: I1206 13:59:25.724172 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdw9k" event={"ID":"906c227f-a713-49d2-acfd-633646b6e0da","Type":"ContainerDied","Data":"19d38017e092899a7cfd0aa5a92e4bd96a09ce9a9759dc303fb55cbd529fceba"} Dec 06 13:59:25 crc kubenswrapper[4706]: I1206 13:59:25.727126 4706 generic.go:334] "Generic (PLEG): container finished" podID="437efad3-ab4d-4e73-86c1-79f182efabbb" containerID="0cc5677112f9507ab5a0ad6e8958e05b3eb0639161748ef868b8cf1178a58fad" exitCode=0 Dec 06 13:59:25 crc kubenswrapper[4706]: I1206 13:59:25.727161 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v9dc2" event={"ID":"437efad3-ab4d-4e73-86c1-79f182efabbb","Type":"ContainerDied","Data":"0cc5677112f9507ab5a0ad6e8958e05b3eb0639161748ef868b8cf1178a58fad"} Dec 06 13:59:25 crc kubenswrapper[4706]: I1206 13:59:25.739874 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kslt9" podStartSLOduration=2.851187301 podStartE2EDuration="57.739858993s" podCreationTimestamp="2025-12-06 13:58:28 +0000 UTC" firstStartedPulling="2025-12-06 13:58:30.179292032 +0000 UTC m=+153.015084889" lastFinishedPulling="2025-12-06 13:59:25.067963714 +0000 UTC m=+207.903756581" observedRunningTime="2025-12-06 13:59:25.737357266 +0000 UTC m=+208.573150153" watchObservedRunningTime="2025-12-06 13:59:25.739858993 +0000 UTC m=+208.575651860" Dec 06 13:59:26 crc kubenswrapper[4706]: I1206 13:59:26.734207 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdw9k" event={"ID":"906c227f-a713-49d2-acfd-633646b6e0da","Type":"ContainerStarted","Data":"2889d570d3639df1831a6543ad6ef2320dbe0c6d06bbce50397a22fc9d2582de"} Dec 06 13:59:26 crc kubenswrapper[4706]: I1206 13:59:26.736456 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v9dc2" event={"ID":"437efad3-ab4d-4e73-86c1-79f182efabbb","Type":"ContainerStarted","Data":"c2afe088d89198040e97fcc95b2b332a26a4ff1cd721ba2df4a96be59079253a"} Dec 06 13:59:26 crc kubenswrapper[4706]: I1206 13:59:26.738220 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjkcq" event={"ID":"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5","Type":"ContainerStarted","Data":"08397677c6ee59183c74d79ad34b5b042c1caba72fb7831df4e72a09c6472615"} Dec 06 13:59:26 crc kubenswrapper[4706]: I1206 13:59:26.739956 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5pm6" event={"ID":"34cf831e-196b-46c2-a2a1-d985cdfd8861","Type":"ContainerStarted","Data":"e192b62048f70495e09c722bfaf3f7bd4724fe91c292baeb5a3cf18618672522"} Dec 06 13:59:26 crc kubenswrapper[4706]: I1206 13:59:26.772135 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wdw9k" podStartSLOduration=2.786551603 podStartE2EDuration="56.772116497s" podCreationTimestamp="2025-12-06 13:58:30 +0000 UTC" firstStartedPulling="2025-12-06 13:58:32.256978842 +0000 UTC m=+155.092771699" lastFinishedPulling="2025-12-06 13:59:26.242543736 +0000 UTC m=+209.078336593" observedRunningTime="2025-12-06 13:59:26.755233727 +0000 UTC m=+209.591026584" watchObservedRunningTime="2025-12-06 13:59:26.772116497 +0000 UTC m=+209.607909354" Dec 06 13:59:26 crc kubenswrapper[4706]: I1206 13:59:26.774706 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v9dc2" podStartSLOduration=2.523263184 podStartE2EDuration="55.774701106s" podCreationTimestamp="2025-12-06 13:58:31 +0000 UTC" firstStartedPulling="2025-12-06 13:58:33.247801024 +0000 UTC m=+156.083593871" lastFinishedPulling="2025-12-06 13:59:26.499238926 +0000 UTC m=+209.335031793" observedRunningTime="2025-12-06 13:59:26.770870734 +0000 UTC m=+209.606663591" watchObservedRunningTime="2025-12-06 13:59:26.774701106 +0000 UTC m=+209.610493963" Dec 06 13:59:27 crc kubenswrapper[4706]: I1206 13:59:27.745938 4706 generic.go:334] "Generic (PLEG): container finished" podID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" containerID="08397677c6ee59183c74d79ad34b5b042c1caba72fb7831df4e72a09c6472615" exitCode=0 Dec 06 13:59:27 crc kubenswrapper[4706]: I1206 13:59:27.746005 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjkcq" event={"ID":"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5","Type":"ContainerDied","Data":"08397677c6ee59183c74d79ad34b5b042c1caba72fb7831df4e72a09c6472615"} Dec 06 13:59:27 crc kubenswrapper[4706]: I1206 13:59:27.747907 4706 generic.go:334] "Generic (PLEG): container finished" podID="34cf831e-196b-46c2-a2a1-d985cdfd8861" containerID="e192b62048f70495e09c722bfaf3f7bd4724fe91c292baeb5a3cf18618672522" exitCode=0 Dec 06 13:59:27 crc kubenswrapper[4706]: I1206 13:59:27.747958 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5pm6" event={"ID":"34cf831e-196b-46c2-a2a1-d985cdfd8861","Type":"ContainerDied","Data":"e192b62048f70495e09c722bfaf3f7bd4724fe91c292baeb5a3cf18618672522"} Dec 06 13:59:28 crc kubenswrapper[4706]: I1206 13:59:28.887267 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kslt9" Dec 06 13:59:28 crc kubenswrapper[4706]: I1206 13:59:28.887608 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kslt9" Dec 06 13:59:28 crc kubenswrapper[4706]: I1206 13:59:28.938993 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kslt9" Dec 06 13:59:29 crc kubenswrapper[4706]: I1206 13:59:29.495379 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gx45h" Dec 06 13:59:30 crc kubenswrapper[4706]: I1206 13:59:30.682026 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wdw9k" Dec 06 13:59:30 crc kubenswrapper[4706]: I1206 13:59:30.692411 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wdw9k" Dec 06 13:59:30 crc kubenswrapper[4706]: I1206 13:59:30.766046 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wdw9k" Dec 06 13:59:31 crc kubenswrapper[4706]: I1206 13:59:31.087762 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sm9zj" Dec 06 13:59:31 crc kubenswrapper[4706]: I1206 13:59:31.088481 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sm9zj" Dec 06 13:59:31 crc kubenswrapper[4706]: I1206 13:59:31.135205 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sm9zj" Dec 06 13:59:31 crc kubenswrapper[4706]: I1206 13:59:31.149584 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gx45h"] Dec 06 13:59:31 crc kubenswrapper[4706]: I1206 13:59:31.149874 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gx45h" podUID="06e9400a-9432-4f75-af53-78b3b3facd7a" containerName="registry-server" containerID="cri-o://8285d11499d6bbaad4a4a9e05fcb3e203134e82112e069839dac6a72c5be5e16" gracePeriod=2 Dec 06 13:59:31 crc kubenswrapper[4706]: I1206 13:59:31.824790 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wdw9k" Dec 06 13:59:31 crc kubenswrapper[4706]: I1206 13:59:31.851488 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sm9zj" Dec 06 13:59:32 crc kubenswrapper[4706]: I1206 13:59:32.299702 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v9dc2" Dec 06 13:59:32 crc kubenswrapper[4706]: I1206 13:59:32.299745 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v9dc2" Dec 06 13:59:33 crc kubenswrapper[4706]: I1206 13:59:33.339703 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v9dc2" podUID="437efad3-ab4d-4e73-86c1-79f182efabbb" containerName="registry-server" probeResult="failure" output=< Dec 06 13:59:33 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 06 13:59:33 crc kubenswrapper[4706]: > Dec 06 13:59:33 crc kubenswrapper[4706]: I1206 13:59:33.791183 4706 generic.go:334] "Generic (PLEG): container finished" podID="06e9400a-9432-4f75-af53-78b3b3facd7a" containerID="8285d11499d6bbaad4a4a9e05fcb3e203134e82112e069839dac6a72c5be5e16" exitCode=0 Dec 06 13:59:33 crc kubenswrapper[4706]: I1206 13:59:33.791296 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gx45h" event={"ID":"06e9400a-9432-4f75-af53-78b3b3facd7a","Type":"ContainerDied","Data":"8285d11499d6bbaad4a4a9e05fcb3e203134e82112e069839dac6a72c5be5e16"} Dec 06 13:59:34 crc kubenswrapper[4706]: I1206 13:59:34.551061 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sm9zj"] Dec 06 13:59:34 crc kubenswrapper[4706]: I1206 13:59:34.797583 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sm9zj" podUID="ebf6ef5b-30f8-4959-b07a-9b0512b19710" containerName="registry-server" containerID="cri-o://8a3c783d310d954db81fa89ba90c77cb5f46dfc7b15ecc13a4425c31c1f841d9" gracePeriod=2 Dec 06 13:59:34 crc kubenswrapper[4706]: I1206 13:59:34.929926 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gx45h" Dec 06 13:59:34 crc kubenswrapper[4706]: I1206 13:59:34.961973 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2l68w\" (UniqueName: \"kubernetes.io/projected/06e9400a-9432-4f75-af53-78b3b3facd7a-kube-api-access-2l68w\") pod \"06e9400a-9432-4f75-af53-78b3b3facd7a\" (UID: \"06e9400a-9432-4f75-af53-78b3b3facd7a\") " Dec 06 13:59:34 crc kubenswrapper[4706]: I1206 13:59:34.962054 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06e9400a-9432-4f75-af53-78b3b3facd7a-catalog-content\") pod \"06e9400a-9432-4f75-af53-78b3b3facd7a\" (UID: \"06e9400a-9432-4f75-af53-78b3b3facd7a\") " Dec 06 13:59:34 crc kubenswrapper[4706]: I1206 13:59:34.962171 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06e9400a-9432-4f75-af53-78b3b3facd7a-utilities\") pod \"06e9400a-9432-4f75-af53-78b3b3facd7a\" (UID: \"06e9400a-9432-4f75-af53-78b3b3facd7a\") " Dec 06 13:59:34 crc kubenswrapper[4706]: I1206 13:59:34.962953 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06e9400a-9432-4f75-af53-78b3b3facd7a-utilities" (OuterVolumeSpecName: "utilities") pod "06e9400a-9432-4f75-af53-78b3b3facd7a" (UID: "06e9400a-9432-4f75-af53-78b3b3facd7a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 13:59:34 crc kubenswrapper[4706]: I1206 13:59:34.968827 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06e9400a-9432-4f75-af53-78b3b3facd7a-kube-api-access-2l68w" (OuterVolumeSpecName: "kube-api-access-2l68w") pod "06e9400a-9432-4f75-af53-78b3b3facd7a" (UID: "06e9400a-9432-4f75-af53-78b3b3facd7a"). InnerVolumeSpecName "kube-api-access-2l68w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:59:35 crc kubenswrapper[4706]: I1206 13:59:35.039891 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06e9400a-9432-4f75-af53-78b3b3facd7a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06e9400a-9432-4f75-af53-78b3b3facd7a" (UID: "06e9400a-9432-4f75-af53-78b3b3facd7a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 13:59:35 crc kubenswrapper[4706]: I1206 13:59:35.064017 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06e9400a-9432-4f75-af53-78b3b3facd7a-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:35 crc kubenswrapper[4706]: I1206 13:59:35.064201 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2l68w\" (UniqueName: \"kubernetes.io/projected/06e9400a-9432-4f75-af53-78b3b3facd7a-kube-api-access-2l68w\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:35 crc kubenswrapper[4706]: I1206 13:59:35.064274 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06e9400a-9432-4f75-af53-78b3b3facd7a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:35 crc kubenswrapper[4706]: I1206 13:59:35.807042 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gx45h" event={"ID":"06e9400a-9432-4f75-af53-78b3b3facd7a","Type":"ContainerDied","Data":"cb2829434fb2137e63a72f6f9fde965d2074d9bb96386957b60a6d708a2704d2"} Dec 06 13:59:35 crc kubenswrapper[4706]: I1206 13:59:35.807586 4706 scope.go:117] "RemoveContainer" containerID="8285d11499d6bbaad4a4a9e05fcb3e203134e82112e069839dac6a72c5be5e16" Dec 06 13:59:35 crc kubenswrapper[4706]: I1206 13:59:35.807083 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gx45h" Dec 06 13:59:35 crc kubenswrapper[4706]: I1206 13:59:35.837865 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gx45h"] Dec 06 13:59:35 crc kubenswrapper[4706]: I1206 13:59:35.841073 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gx45h"] Dec 06 13:59:36 crc kubenswrapper[4706]: I1206 13:59:36.513667 4706 scope.go:117] "RemoveContainer" containerID="6dfefcbf013ff43a6bfcc38e1ba8427c10e0249435ade86f54bd4c27b8c88b14" Dec 06 13:59:36 crc kubenswrapper[4706]: I1206 13:59:36.821618 4706 generic.go:334] "Generic (PLEG): container finished" podID="ebf6ef5b-30f8-4959-b07a-9b0512b19710" containerID="8a3c783d310d954db81fa89ba90c77cb5f46dfc7b15ecc13a4425c31c1f841d9" exitCode=0 Dec 06 13:59:36 crc kubenswrapper[4706]: I1206 13:59:36.821682 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sm9zj" event={"ID":"ebf6ef5b-30f8-4959-b07a-9b0512b19710","Type":"ContainerDied","Data":"8a3c783d310d954db81fa89ba90c77cb5f46dfc7b15ecc13a4425c31c1f841d9"} Dec 06 13:59:37 crc kubenswrapper[4706]: I1206 13:59:37.524240 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06e9400a-9432-4f75-af53-78b3b3facd7a" path="/var/lib/kubelet/pods/06e9400a-9432-4f75-af53-78b3b3facd7a/volumes" Dec 06 13:59:37 crc kubenswrapper[4706]: I1206 13:59:37.684671 4706 scope.go:117] "RemoveContainer" containerID="114116940b690d4be2b7e173f264304b324950eb3a9911e95a70f5f8e45b8dfc" Dec 06 13:59:37 crc kubenswrapper[4706]: I1206 13:59:37.726652 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sm9zj" Dec 06 13:59:37 crc kubenswrapper[4706]: I1206 13:59:37.807009 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcp94\" (UniqueName: \"kubernetes.io/projected/ebf6ef5b-30f8-4959-b07a-9b0512b19710-kube-api-access-hcp94\") pod \"ebf6ef5b-30f8-4959-b07a-9b0512b19710\" (UID: \"ebf6ef5b-30f8-4959-b07a-9b0512b19710\") " Dec 06 13:59:37 crc kubenswrapper[4706]: I1206 13:59:37.807084 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebf6ef5b-30f8-4959-b07a-9b0512b19710-catalog-content\") pod \"ebf6ef5b-30f8-4959-b07a-9b0512b19710\" (UID: \"ebf6ef5b-30f8-4959-b07a-9b0512b19710\") " Dec 06 13:59:37 crc kubenswrapper[4706]: I1206 13:59:37.807123 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebf6ef5b-30f8-4959-b07a-9b0512b19710-utilities\") pod \"ebf6ef5b-30f8-4959-b07a-9b0512b19710\" (UID: \"ebf6ef5b-30f8-4959-b07a-9b0512b19710\") " Dec 06 13:59:37 crc kubenswrapper[4706]: I1206 13:59:37.808488 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebf6ef5b-30f8-4959-b07a-9b0512b19710-utilities" (OuterVolumeSpecName: "utilities") pod "ebf6ef5b-30f8-4959-b07a-9b0512b19710" (UID: "ebf6ef5b-30f8-4959-b07a-9b0512b19710"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 13:59:37 crc kubenswrapper[4706]: I1206 13:59:37.816959 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebf6ef5b-30f8-4959-b07a-9b0512b19710-kube-api-access-hcp94" (OuterVolumeSpecName: "kube-api-access-hcp94") pod "ebf6ef5b-30f8-4959-b07a-9b0512b19710" (UID: "ebf6ef5b-30f8-4959-b07a-9b0512b19710"). InnerVolumeSpecName "kube-api-access-hcp94". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:59:37 crc kubenswrapper[4706]: I1206 13:59:37.832854 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sm9zj" event={"ID":"ebf6ef5b-30f8-4959-b07a-9b0512b19710","Type":"ContainerDied","Data":"066a4b6604da17b9be68ab47f5d6a3804b1cab336abde94d43f5beb57f6355d7"} Dec 06 13:59:37 crc kubenswrapper[4706]: I1206 13:59:37.832936 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sm9zj" Dec 06 13:59:37 crc kubenswrapper[4706]: I1206 13:59:37.837999 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebf6ef5b-30f8-4959-b07a-9b0512b19710-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ebf6ef5b-30f8-4959-b07a-9b0512b19710" (UID: "ebf6ef5b-30f8-4959-b07a-9b0512b19710"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 13:59:37 crc kubenswrapper[4706]: I1206 13:59:37.909081 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebf6ef5b-30f8-4959-b07a-9b0512b19710-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:37 crc kubenswrapper[4706]: I1206 13:59:37.909124 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcp94\" (UniqueName: \"kubernetes.io/projected/ebf6ef5b-30f8-4959-b07a-9b0512b19710-kube-api-access-hcp94\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:37 crc kubenswrapper[4706]: I1206 13:59:37.909142 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebf6ef5b-30f8-4959-b07a-9b0512b19710-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:38 crc kubenswrapper[4706]: I1206 13:59:38.174301 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sm9zj"] Dec 06 13:59:38 crc kubenswrapper[4706]: I1206 13:59:38.181801 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sm9zj"] Dec 06 13:59:38 crc kubenswrapper[4706]: I1206 13:59:38.958168 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kslt9" Dec 06 13:59:39 crc kubenswrapper[4706]: I1206 13:59:39.518732 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebf6ef5b-30f8-4959-b07a-9b0512b19710" path="/var/lib/kubelet/pods/ebf6ef5b-30f8-4959-b07a-9b0512b19710/volumes" Dec 06 13:59:40 crc kubenswrapper[4706]: I1206 13:59:40.223715 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" podUID="7898190a-5957-4772-abbb-1bfbf16f4eb5" containerName="oauth-openshift" containerID="cri-o://4b0d06d22caa9b797f856c01a75d8a5de765373ca71b47a6c4c24cff7402ae88" gracePeriod=15 Dec 06 13:59:40 crc kubenswrapper[4706]: I1206 13:59:40.858775 4706 generic.go:334] "Generic (PLEG): container finished" podID="7898190a-5957-4772-abbb-1bfbf16f4eb5" containerID="4b0d06d22caa9b797f856c01a75d8a5de765373ca71b47a6c4c24cff7402ae88" exitCode=0 Dec 06 13:59:40 crc kubenswrapper[4706]: I1206 13:59:40.858821 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" event={"ID":"7898190a-5957-4772-abbb-1bfbf16f4eb5","Type":"ContainerDied","Data":"4b0d06d22caa9b797f856c01a75d8a5de765373ca71b47a6c4c24cff7402ae88"} Dec 06 13:59:41 crc kubenswrapper[4706]: I1206 13:59:41.848336 4706 scope.go:117] "RemoveContainer" containerID="8a3c783d310d954db81fa89ba90c77cb5f46dfc7b15ecc13a4425c31c1f841d9" Dec 06 13:59:42 crc kubenswrapper[4706]: I1206 13:59:42.370742 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v9dc2" Dec 06 13:59:42 crc kubenswrapper[4706]: I1206 13:59:42.431894 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v9dc2" Dec 06 13:59:42 crc kubenswrapper[4706]: I1206 13:59:42.608841 4706 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-l55q6 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.18:6443/healthz\": dial tcp 10.217.0.18:6443: connect: connection refused" start-of-body= Dec 06 13:59:42 crc kubenswrapper[4706]: I1206 13:59:42.608911 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" podUID="7898190a-5957-4772-abbb-1bfbf16f4eb5" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.18:6443/healthz\": dial tcp 10.217.0.18:6443: connect: connection refused" Dec 06 13:59:43 crc kubenswrapper[4706]: I1206 13:59:43.349910 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v9dc2"] Dec 06 13:59:43 crc kubenswrapper[4706]: I1206 13:59:43.880583 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v9dc2" podUID="437efad3-ab4d-4e73-86c1-79f182efabbb" containerName="registry-server" containerID="cri-o://c2afe088d89198040e97fcc95b2b332a26a4ff1cd721ba2df4a96be59079253a" gracePeriod=2 Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.580813 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.581214 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.581276 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.582031 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f"} pod="openshift-machine-config-operator/machine-config-daemon-67v56" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.582167 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" containerID="cri-o://244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f" gracePeriod=600 Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.730673 4706 scope.go:117] "RemoveContainer" containerID="1a8b649e0b36041830affaf539efee5bb01292cf814baecd3da267e883f0326b" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.807094 4706 scope.go:117] "RemoveContainer" containerID="cd7776a885ec55e84f412b74be8186a6444007a15547a7088274c1907e7c98b8" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.833220 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.863356 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-9565f95f5-tjn97"] Dec 06 13:59:44 crc kubenswrapper[4706]: E1206 13:59:44.863669 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e9400a-9432-4f75-af53-78b3b3facd7a" containerName="registry-server" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.863688 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e9400a-9432-4f75-af53-78b3b3facd7a" containerName="registry-server" Dec 06 13:59:44 crc kubenswrapper[4706]: E1206 13:59:44.863701 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7898190a-5957-4772-abbb-1bfbf16f4eb5" containerName="oauth-openshift" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.863730 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="7898190a-5957-4772-abbb-1bfbf16f4eb5" containerName="oauth-openshift" Dec 06 13:59:44 crc kubenswrapper[4706]: E1206 13:59:44.863745 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf6ef5b-30f8-4959-b07a-9b0512b19710" containerName="extract-utilities" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.863754 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf6ef5b-30f8-4959-b07a-9b0512b19710" containerName="extract-utilities" Dec 06 13:59:44 crc kubenswrapper[4706]: E1206 13:59:44.863770 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf6ef5b-30f8-4959-b07a-9b0512b19710" containerName="extract-content" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.863779 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf6ef5b-30f8-4959-b07a-9b0512b19710" containerName="extract-content" Dec 06 13:59:44 crc kubenswrapper[4706]: E1206 13:59:44.863812 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e9400a-9432-4f75-af53-78b3b3facd7a" containerName="extract-content" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.863820 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e9400a-9432-4f75-af53-78b3b3facd7a" containerName="extract-content" Dec 06 13:59:44 crc kubenswrapper[4706]: E1206 13:59:44.863830 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf6ef5b-30f8-4959-b07a-9b0512b19710" containerName="registry-server" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.863838 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf6ef5b-30f8-4959-b07a-9b0512b19710" containerName="registry-server" Dec 06 13:59:44 crc kubenswrapper[4706]: E1206 13:59:44.863850 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ceeceb1-f929-4022-9c4c-68d608dfdaaa" containerName="pruner" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.863857 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ceeceb1-f929-4022-9c4c-68d608dfdaaa" containerName="pruner" Dec 06 13:59:44 crc kubenswrapper[4706]: E1206 13:59:44.863868 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e9400a-9432-4f75-af53-78b3b3facd7a" containerName="extract-utilities" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.863897 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e9400a-9432-4f75-af53-78b3b3facd7a" containerName="extract-utilities" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.864064 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="7898190a-5957-4772-abbb-1bfbf16f4eb5" containerName="oauth-openshift" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.864082 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="06e9400a-9432-4f75-af53-78b3b3facd7a" containerName="registry-server" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.864092 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ceeceb1-f929-4022-9c4c-68d608dfdaaa" containerName="pruner" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.864102 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebf6ef5b-30f8-4959-b07a-9b0512b19710" containerName="registry-server" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.864745 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.872184 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-9565f95f5-tjn97"] Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.895437 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.895452 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-l55q6" event={"ID":"7898190a-5957-4772-abbb-1bfbf16f4eb5","Type":"ContainerDied","Data":"d5109d0031068713991c9cf3045d11f4a0988baf1963dc4cd68791a06fe6d7ea"} Dec 06 13:59:44 crc kubenswrapper[4706]: I1206 13:59:44.895536 4706 scope.go:117] "RemoveContainer" containerID="4b0d06d22caa9b797f856c01a75d8a5de765373ca71b47a6c4c24cff7402ae88" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.010653 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-idp-0-file-data\") pod \"7898190a-5957-4772-abbb-1bfbf16f4eb5\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.010704 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-service-ca\") pod \"7898190a-5957-4772-abbb-1bfbf16f4eb5\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.010722 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-session\") pod \"7898190a-5957-4772-abbb-1bfbf16f4eb5\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.010740 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-login\") pod \"7898190a-5957-4772-abbb-1bfbf16f4eb5\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.010773 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-provider-selection\") pod \"7898190a-5957-4772-abbb-1bfbf16f4eb5\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.010808 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-cliconfig\") pod \"7898190a-5957-4772-abbb-1bfbf16f4eb5\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.010843 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-audit-policies\") pod \"7898190a-5957-4772-abbb-1bfbf16f4eb5\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.010860 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-error\") pod \"7898190a-5957-4772-abbb-1bfbf16f4eb5\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.010877 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bc9m\" (UniqueName: \"kubernetes.io/projected/7898190a-5957-4772-abbb-1bfbf16f4eb5-kube-api-access-9bc9m\") pod \"7898190a-5957-4772-abbb-1bfbf16f4eb5\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.010896 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-router-certs\") pod \"7898190a-5957-4772-abbb-1bfbf16f4eb5\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.010918 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-trusted-ca-bundle\") pod \"7898190a-5957-4772-abbb-1bfbf16f4eb5\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.010931 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7898190a-5957-4772-abbb-1bfbf16f4eb5-audit-dir\") pod \"7898190a-5957-4772-abbb-1bfbf16f4eb5\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.010956 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-serving-cert\") pod \"7898190a-5957-4772-abbb-1bfbf16f4eb5\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.010989 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-ocp-branding-template\") pod \"7898190a-5957-4772-abbb-1bfbf16f4eb5\" (UID: \"7898190a-5957-4772-abbb-1bfbf16f4eb5\") " Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.011126 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/00317e94-33c6-48ec-bd52-1809d6499cae-audit-policies\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.011154 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.011169 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.011189 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.011214 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-service-ca\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.011233 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.011261 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.011278 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.011296 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-user-template-login\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.011335 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/00317e94-33c6-48ec-bd52-1809d6499cae-audit-dir\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.011352 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4hqz\" (UniqueName: \"kubernetes.io/projected/00317e94-33c6-48ec-bd52-1809d6499cae-kube-api-access-x4hqz\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.011373 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-session\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.011390 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-user-template-error\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.011415 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-router-certs\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.012472 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "7898190a-5957-4772-abbb-1bfbf16f4eb5" (UID: "7898190a-5957-4772-abbb-1bfbf16f4eb5"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.012545 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "7898190a-5957-4772-abbb-1bfbf16f4eb5" (UID: "7898190a-5957-4772-abbb-1bfbf16f4eb5"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.012565 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7898190a-5957-4772-abbb-1bfbf16f4eb5-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "7898190a-5957-4772-abbb-1bfbf16f4eb5" (UID: "7898190a-5957-4772-abbb-1bfbf16f4eb5"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.013162 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "7898190a-5957-4772-abbb-1bfbf16f4eb5" (UID: "7898190a-5957-4772-abbb-1bfbf16f4eb5"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.013410 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "7898190a-5957-4772-abbb-1bfbf16f4eb5" (UID: "7898190a-5957-4772-abbb-1bfbf16f4eb5"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.016616 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "7898190a-5957-4772-abbb-1bfbf16f4eb5" (UID: "7898190a-5957-4772-abbb-1bfbf16f4eb5"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.017032 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "7898190a-5957-4772-abbb-1bfbf16f4eb5" (UID: "7898190a-5957-4772-abbb-1bfbf16f4eb5"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.017484 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "7898190a-5957-4772-abbb-1bfbf16f4eb5" (UID: "7898190a-5957-4772-abbb-1bfbf16f4eb5"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.018378 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "7898190a-5957-4772-abbb-1bfbf16f4eb5" (UID: "7898190a-5957-4772-abbb-1bfbf16f4eb5"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.018748 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "7898190a-5957-4772-abbb-1bfbf16f4eb5" (UID: "7898190a-5957-4772-abbb-1bfbf16f4eb5"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.018860 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "7898190a-5957-4772-abbb-1bfbf16f4eb5" (UID: "7898190a-5957-4772-abbb-1bfbf16f4eb5"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.019137 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "7898190a-5957-4772-abbb-1bfbf16f4eb5" (UID: "7898190a-5957-4772-abbb-1bfbf16f4eb5"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.019279 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "7898190a-5957-4772-abbb-1bfbf16f4eb5" (UID: "7898190a-5957-4772-abbb-1bfbf16f4eb5"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.020530 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7898190a-5957-4772-abbb-1bfbf16f4eb5-kube-api-access-9bc9m" (OuterVolumeSpecName: "kube-api-access-9bc9m") pod "7898190a-5957-4772-abbb-1bfbf16f4eb5" (UID: "7898190a-5957-4772-abbb-1bfbf16f4eb5"). InnerVolumeSpecName "kube-api-access-9bc9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113014 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-router-certs\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113088 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/00317e94-33c6-48ec-bd52-1809d6499cae-audit-policies\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113129 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113152 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113180 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113208 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-service-ca\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113234 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113271 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113351 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113380 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-user-template-login\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113417 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/00317e94-33c6-48ec-bd52-1809d6499cae-audit-dir\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113441 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4hqz\" (UniqueName: \"kubernetes.io/projected/00317e94-33c6-48ec-bd52-1809d6499cae-kube-api-access-x4hqz\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113467 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-session\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113490 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-user-template-error\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113550 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113566 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113579 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113591 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113604 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113618 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113631 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113644 4706 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113658 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113671 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bc9m\" (UniqueName: \"kubernetes.io/projected/7898190a-5957-4772-abbb-1bfbf16f4eb5-kube-api-access-9bc9m\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113683 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113695 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113708 4706 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7898190a-5957-4772-abbb-1bfbf16f4eb5-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.113720 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7898190a-5957-4772-abbb-1bfbf16f4eb5-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.115403 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/00317e94-33c6-48ec-bd52-1809d6499cae-audit-dir\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.116635 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.117543 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/00317e94-33c6-48ec-bd52-1809d6499cae-audit-policies\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.118551 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-user-template-error\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.119801 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-service-ca\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.119953 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.121108 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.122730 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-session\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.123202 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-user-template-login\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.123204 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.123811 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.124645 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.125132 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/00317e94-33c6-48ec-bd52-1809d6499cae-v4-0-config-system-router-certs\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.134429 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4hqz\" (UniqueName: \"kubernetes.io/projected/00317e94-33c6-48ec-bd52-1809d6499cae-kube-api-access-x4hqz\") pod \"oauth-openshift-9565f95f5-tjn97\" (UID: \"00317e94-33c6-48ec-bd52-1809d6499cae\") " pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.229643 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l55q6"] Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.233714 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l55q6"] Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.239157 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.520222 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7898190a-5957-4772-abbb-1bfbf16f4eb5" path="/var/lib/kubelet/pods/7898190a-5957-4772-abbb-1bfbf16f4eb5/volumes" Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.742219 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-9565f95f5-tjn97"] Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.921733 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5pm6" event={"ID":"34cf831e-196b-46c2-a2a1-d985cdfd8861","Type":"ContainerStarted","Data":"6f740d3590d0a3cf15468d637d45fb68f3ecd5bdcf66d802d6cadc1bf62a80bc"} Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.927732 4706 generic.go:334] "Generic (PLEG): container finished" podID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerID="244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f" exitCode=0 Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.927821 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerDied","Data":"244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f"} Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.929958 4706 generic.go:334] "Generic (PLEG): container finished" podID="437efad3-ab4d-4e73-86c1-79f182efabbb" containerID="c2afe088d89198040e97fcc95b2b332a26a4ff1cd721ba2df4a96be59079253a" exitCode=0 Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.929997 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v9dc2" event={"ID":"437efad3-ab4d-4e73-86c1-79f182efabbb","Type":"ContainerDied","Data":"c2afe088d89198040e97fcc95b2b332a26a4ff1cd721ba2df4a96be59079253a"} Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.930979 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" event={"ID":"00317e94-33c6-48ec-bd52-1809d6499cae","Type":"ContainerStarted","Data":"f64585f7a1dbdce7ef5f0814d91502d1427f7d19a911989a6cca91d252dd5dc4"} Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.932566 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjkcq" event={"ID":"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5","Type":"ContainerStarted","Data":"0e79497c67181657b28b5995c7609d75815eb1bddd2ee6261744d3cfdcd41f01"} Dec 06 13:59:45 crc kubenswrapper[4706]: I1206 13:59:45.933608 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zhlvc" event={"ID":"d12f4107-0d89-4d3c-958f-7355fbd27575","Type":"ContainerStarted","Data":"8f77e7cfa55d0de9299ca6f6184f7d4a3cb8ee885a51d505f2a8fef78438c9dc"} Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.352474 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v9dc2" Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.529372 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzsv2\" (UniqueName: \"kubernetes.io/projected/437efad3-ab4d-4e73-86c1-79f182efabbb-kube-api-access-nzsv2\") pod \"437efad3-ab4d-4e73-86c1-79f182efabbb\" (UID: \"437efad3-ab4d-4e73-86c1-79f182efabbb\") " Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.529478 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/437efad3-ab4d-4e73-86c1-79f182efabbb-catalog-content\") pod \"437efad3-ab4d-4e73-86c1-79f182efabbb\" (UID: \"437efad3-ab4d-4e73-86c1-79f182efabbb\") " Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.529659 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/437efad3-ab4d-4e73-86c1-79f182efabbb-utilities\") pod \"437efad3-ab4d-4e73-86c1-79f182efabbb\" (UID: \"437efad3-ab4d-4e73-86c1-79f182efabbb\") " Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.530707 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/437efad3-ab4d-4e73-86c1-79f182efabbb-utilities" (OuterVolumeSpecName: "utilities") pod "437efad3-ab4d-4e73-86c1-79f182efabbb" (UID: "437efad3-ab4d-4e73-86c1-79f182efabbb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.534867 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/437efad3-ab4d-4e73-86c1-79f182efabbb-kube-api-access-nzsv2" (OuterVolumeSpecName: "kube-api-access-nzsv2") pod "437efad3-ab4d-4e73-86c1-79f182efabbb" (UID: "437efad3-ab4d-4e73-86c1-79f182efabbb"). InnerVolumeSpecName "kube-api-access-nzsv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.631293 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/437efad3-ab4d-4e73-86c1-79f182efabbb-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.631847 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzsv2\" (UniqueName: \"kubernetes.io/projected/437efad3-ab4d-4e73-86c1-79f182efabbb-kube-api-access-nzsv2\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.631412 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/437efad3-ab4d-4e73-86c1-79f182efabbb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "437efad3-ab4d-4e73-86c1-79f182efabbb" (UID: "437efad3-ab4d-4e73-86c1-79f182efabbb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.733229 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/437efad3-ab4d-4e73-86c1-79f182efabbb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.940526 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" event={"ID":"00317e94-33c6-48ec-bd52-1809d6499cae","Type":"ContainerStarted","Data":"c2287a4c25a47d6321c7f486ba939365ce24cff533fff50491f3801d9d324f4a"} Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.941922 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.943894 4706 generic.go:334] "Generic (PLEG): container finished" podID="d12f4107-0d89-4d3c-958f-7355fbd27575" containerID="8f77e7cfa55d0de9299ca6f6184f7d4a3cb8ee885a51d505f2a8fef78438c9dc" exitCode=0 Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.944023 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zhlvc" event={"ID":"d12f4107-0d89-4d3c-958f-7355fbd27575","Type":"ContainerDied","Data":"8f77e7cfa55d0de9299ca6f6184f7d4a3cb8ee885a51d505f2a8fef78438c9dc"} Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.951739 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.951777 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerStarted","Data":"067a72c2e70239f7868cd95da2f6bc3f2b176f2e63176fd1a3892a751afb382e"} Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.954410 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v9dc2" event={"ID":"437efad3-ab4d-4e73-86c1-79f182efabbb","Type":"ContainerDied","Data":"a57dfaae493e5dbc17ecd71d8d9a55bd5e93a1c0e0a1a0491fe0a83826f9a1b9"} Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.954462 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v9dc2" Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.954488 4706 scope.go:117] "RemoveContainer" containerID="c2afe088d89198040e97fcc95b2b332a26a4ff1cd721ba2df4a96be59079253a" Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.972504 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-9565f95f5-tjn97" podStartSLOduration=31.972483124 podStartE2EDuration="31.972483124s" podCreationTimestamp="2025-12-06 13:59:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 13:59:46.969977948 +0000 UTC m=+229.805770815" watchObservedRunningTime="2025-12-06 13:59:46.972483124 +0000 UTC m=+229.808275991" Dec 06 13:59:46 crc kubenswrapper[4706]: I1206 13:59:46.989497 4706 scope.go:117] "RemoveContainer" containerID="0cc5677112f9507ab5a0ad6e8958e05b3eb0639161748ef868b8cf1178a58fad" Dec 06 13:59:47 crc kubenswrapper[4706]: I1206 13:59:47.040790 4706 scope.go:117] "RemoveContainer" containerID="ee7db2a27d3794da2faf8839a0d77e17852104bc4bbb231bcedf2e4772f91392" Dec 06 13:59:47 crc kubenswrapper[4706]: I1206 13:59:47.064830 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x5pm6" podStartSLOduration=4.575833428 podStartE2EDuration="1m16.064812238s" podCreationTimestamp="2025-12-06 13:58:31 +0000 UTC" firstStartedPulling="2025-12-06 13:58:33.243197815 +0000 UTC m=+156.078990672" lastFinishedPulling="2025-12-06 13:59:44.732176595 +0000 UTC m=+227.567969482" observedRunningTime="2025-12-06 13:59:47.058670424 +0000 UTC m=+229.894463281" watchObservedRunningTime="2025-12-06 13:59:47.064812238 +0000 UTC m=+229.900605095" Dec 06 13:59:47 crc kubenswrapper[4706]: I1206 13:59:47.103409 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hjkcq" podStartSLOduration=7.412473263 podStartE2EDuration="1m19.103392948s" podCreationTimestamp="2025-12-06 13:58:28 +0000 UTC" firstStartedPulling="2025-12-06 13:58:30.166543062 +0000 UTC m=+153.002335919" lastFinishedPulling="2025-12-06 13:59:41.857462717 +0000 UTC m=+224.693255604" observedRunningTime="2025-12-06 13:59:47.095453426 +0000 UTC m=+229.931246283" watchObservedRunningTime="2025-12-06 13:59:47.103392948 +0000 UTC m=+229.939185805" Dec 06 13:59:47 crc kubenswrapper[4706]: I1206 13:59:47.110767 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v9dc2"] Dec 06 13:59:47 crc kubenswrapper[4706]: I1206 13:59:47.114360 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-v9dc2"] Dec 06 13:59:47 crc kubenswrapper[4706]: I1206 13:59:47.521809 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="437efad3-ab4d-4e73-86c1-79f182efabbb" path="/var/lib/kubelet/pods/437efad3-ab4d-4e73-86c1-79f182efabbb/volumes" Dec 06 13:59:48 crc kubenswrapper[4706]: I1206 13:59:48.734447 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hjkcq" Dec 06 13:59:48 crc kubenswrapper[4706]: I1206 13:59:48.734549 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hjkcq" Dec 06 13:59:48 crc kubenswrapper[4706]: I1206 13:59:48.803708 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hjkcq" Dec 06 13:59:48 crc kubenswrapper[4706]: I1206 13:59:48.972677 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zhlvc" event={"ID":"d12f4107-0d89-4d3c-958f-7355fbd27575","Type":"ContainerStarted","Data":"e3ff9ae87bec78ac5f3199a1b0baa2fbbcf7b3f5f0a3bd99d9ae29b94e2c2650"} Dec 06 13:59:49 crc kubenswrapper[4706]: I1206 13:59:49.007543 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zhlvc" podStartSLOduration=3.53823236 podStartE2EDuration="1m21.007505746s" podCreationTimestamp="2025-12-06 13:58:28 +0000 UTC" firstStartedPulling="2025-12-06 13:58:30.170548826 +0000 UTC m=+153.006341683" lastFinishedPulling="2025-12-06 13:59:47.639822212 +0000 UTC m=+230.475615069" observedRunningTime="2025-12-06 13:59:49.003580281 +0000 UTC m=+231.839373168" watchObservedRunningTime="2025-12-06 13:59:49.007505746 +0000 UTC m=+231.843298643" Dec 06 13:59:49 crc kubenswrapper[4706]: I1206 13:59:49.114916 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zhlvc" Dec 06 13:59:49 crc kubenswrapper[4706]: I1206 13:59:49.115097 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zhlvc" Dec 06 13:59:50 crc kubenswrapper[4706]: I1206 13:59:50.199804 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-zhlvc" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" containerName="registry-server" probeResult="failure" output=< Dec 06 13:59:50 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 06 13:59:50 crc kubenswrapper[4706]: > Dec 06 13:59:51 crc kubenswrapper[4706]: I1206 13:59:51.893904 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x5pm6" Dec 06 13:59:51 crc kubenswrapper[4706]: I1206 13:59:51.894168 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x5pm6" Dec 06 13:59:52 crc kubenswrapper[4706]: I1206 13:59:52.951091 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x5pm6" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" containerName="registry-server" probeResult="failure" output=< Dec 06 13:59:52 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Dec 06 13:59:52 crc kubenswrapper[4706]: > Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.488796 4706 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.489768 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788" gracePeriod=15 Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.489834 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42" gracePeriod=15 Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.489809 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad" gracePeriod=15 Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.489786 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05" gracePeriod=15 Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.490373 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0" gracePeriod=15 Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.496150 4706 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 13:59:53 crc kubenswrapper[4706]: E1206 13:59:53.496763 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.496802 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 13:59:53 crc kubenswrapper[4706]: E1206 13:59:53.496832 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.496849 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 13:59:53 crc kubenswrapper[4706]: E1206 13:59:53.496870 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="437efad3-ab4d-4e73-86c1-79f182efabbb" containerName="registry-server" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.496887 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="437efad3-ab4d-4e73-86c1-79f182efabbb" containerName="registry-server" Dec 06 13:59:53 crc kubenswrapper[4706]: E1206 13:59:53.496911 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="437efad3-ab4d-4e73-86c1-79f182efabbb" containerName="extract-utilities" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.496926 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="437efad3-ab4d-4e73-86c1-79f182efabbb" containerName="extract-utilities" Dec 06 13:59:53 crc kubenswrapper[4706]: E1206 13:59:53.496959 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.496978 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 13:59:53 crc kubenswrapper[4706]: E1206 13:59:53.496996 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.497015 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 13:59:53 crc kubenswrapper[4706]: E1206 13:59:53.497039 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="437efad3-ab4d-4e73-86c1-79f182efabbb" containerName="extract-content" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.497055 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="437efad3-ab4d-4e73-86c1-79f182efabbb" containerName="extract-content" Dec 06 13:59:53 crc kubenswrapper[4706]: E1206 13:59:53.497083 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.497100 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 13:59:53 crc kubenswrapper[4706]: E1206 13:59:53.497121 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.497136 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.497393 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.497427 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.497449 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.497476 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="437efad3-ab4d-4e73-86c1-79f182efabbb" containerName="registry-server" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.497500 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.497517 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 13:59:53 crc kubenswrapper[4706]: E1206 13:59:53.497781 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.497802 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.498021 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.504820 4706 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.508513 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.514160 4706 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.537429 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.537506 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.537630 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.537670 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.537773 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.537813 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.537908 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.537939 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: E1206 13:59:53.590011 4706 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.83:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.639990 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.640042 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.640092 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.640088 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.640110 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.640148 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.640122 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.640209 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.640233 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.640251 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.640268 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.640297 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.640299 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.640354 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.640330 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.640239 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: I1206 13:59:53.890679 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:53 crc kubenswrapper[4706]: W1206 13:59:53.920023 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-d5964b472ddc70287d88eda979f036efda4e6bfc0b127850cb8419b7bb834903 WatchSource:0}: Error finding container d5964b472ddc70287d88eda979f036efda4e6bfc0b127850cb8419b7bb834903: Status 404 returned error can't find the container with id d5964b472ddc70287d88eda979f036efda4e6bfc0b127850cb8419b7bb834903 Dec 06 13:59:53 crc kubenswrapper[4706]: E1206 13:59:53.924295 4706 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.83:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187ea5106bdd0149 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 13:59:53.923518793 +0000 UTC m=+236.759311690,LastTimestamp:2025-12-06 13:59:53.923518793 +0000 UTC m=+236.759311690,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 13:59:54 crc kubenswrapper[4706]: I1206 13:59:54.010144 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"d5964b472ddc70287d88eda979f036efda4e6bfc0b127850cb8419b7bb834903"} Dec 06 13:59:54 crc kubenswrapper[4706]: I1206 13:59:54.014821 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 13:59:54 crc kubenswrapper[4706]: I1206 13:59:54.017498 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 13:59:54 crc kubenswrapper[4706]: I1206 13:59:54.018870 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad" exitCode=0 Dec 06 13:59:54 crc kubenswrapper[4706]: I1206 13:59:54.018908 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42" exitCode=0 Dec 06 13:59:54 crc kubenswrapper[4706]: I1206 13:59:54.018924 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05" exitCode=0 Dec 06 13:59:54 crc kubenswrapper[4706]: I1206 13:59:54.018939 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0" exitCode=2 Dec 06 13:59:54 crc kubenswrapper[4706]: I1206 13:59:54.018929 4706 scope.go:117] "RemoveContainer" containerID="09a347a033f5707b7fe1cc4463f71f72ba593eaf218aa9edddd71dca731436a3" Dec 06 13:59:54 crc kubenswrapper[4706]: I1206 13:59:54.021692 4706 generic.go:334] "Generic (PLEG): container finished" podID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" containerID="c671b8c8691a77d80e9cd9d6fcce6bef62c242089e507629952c27a68739df84" exitCode=0 Dec 06 13:59:54 crc kubenswrapper[4706]: I1206 13:59:54.021732 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"40d27479-a513-43d4-8ad7-9a4daa8a29d7","Type":"ContainerDied","Data":"c671b8c8691a77d80e9cd9d6fcce6bef62c242089e507629952c27a68739df84"} Dec 06 13:59:54 crc kubenswrapper[4706]: I1206 13:59:54.022368 4706 status_manager.go:851] "Failed to get status for pod" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.030808 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"1aa02db026a29f6ba7f04e7271a176bb36d434dd04cefb0511151b5520f6579b"} Dec 06 13:59:55 crc kubenswrapper[4706]: E1206 13:59:55.031855 4706 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.83:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.031859 4706 status_manager.go:851] "Failed to get status for pod" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.035714 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.616958 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.620004 4706 status_manager.go:851] "Failed to get status for pod" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.798551 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/40d27479-a513-43d4-8ad7-9a4daa8a29d7-var-lock\") pod \"40d27479-a513-43d4-8ad7-9a4daa8a29d7\" (UID: \"40d27479-a513-43d4-8ad7-9a4daa8a29d7\") " Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.798636 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/40d27479-a513-43d4-8ad7-9a4daa8a29d7-kubelet-dir\") pod \"40d27479-a513-43d4-8ad7-9a4daa8a29d7\" (UID: \"40d27479-a513-43d4-8ad7-9a4daa8a29d7\") " Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.798692 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/40d27479-a513-43d4-8ad7-9a4daa8a29d7-kube-api-access\") pod \"40d27479-a513-43d4-8ad7-9a4daa8a29d7\" (UID: \"40d27479-a513-43d4-8ad7-9a4daa8a29d7\") " Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.798866 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40d27479-a513-43d4-8ad7-9a4daa8a29d7-var-lock" (OuterVolumeSpecName: "var-lock") pod "40d27479-a513-43d4-8ad7-9a4daa8a29d7" (UID: "40d27479-a513-43d4-8ad7-9a4daa8a29d7"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.798874 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40d27479-a513-43d4-8ad7-9a4daa8a29d7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "40d27479-a513-43d4-8ad7-9a4daa8a29d7" (UID: "40d27479-a513-43d4-8ad7-9a4daa8a29d7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.799020 4706 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/40d27479-a513-43d4-8ad7-9a4daa8a29d7-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.799045 4706 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/40d27479-a513-43d4-8ad7-9a4daa8a29d7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.805200 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40d27479-a513-43d4-8ad7-9a4daa8a29d7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "40d27479-a513-43d4-8ad7-9a4daa8a29d7" (UID: "40d27479-a513-43d4-8ad7-9a4daa8a29d7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.901338 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/40d27479-a513-43d4-8ad7-9a4daa8a29d7-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.905914 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.907227 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.908186 4706 status_manager.go:851] "Failed to get status for pod" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:55 crc kubenswrapper[4706]: I1206 13:59:55.908808 4706 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.002559 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.002645 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.002712 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.002921 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.002961 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.002932 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.050481 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.051969 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788" exitCode=0 Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.052153 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.052156 4706 scope.go:117] "RemoveContainer" containerID="914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.054954 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"40d27479-a513-43d4-8ad7-9a4daa8a29d7","Type":"ContainerDied","Data":"f2f91e9bce6f725c95b4835dc183f3f7e0c5179be7b62ed525d943a415873ba3"} Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.055024 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.055048 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2f91e9bce6f725c95b4835dc183f3f7e0c5179be7b62ed525d943a415873ba3" Dec 06 13:59:56 crc kubenswrapper[4706]: E1206 13:59:56.055828 4706 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.83:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.081262 4706 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.081996 4706 scope.go:117] "RemoveContainer" containerID="832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.082339 4706 status_manager.go:851] "Failed to get status for pod" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.086639 4706 status_manager.go:851] "Failed to get status for pod" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.087173 4706 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.102296 4706 scope.go:117] "RemoveContainer" containerID="603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.104355 4706 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.104398 4706 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.104415 4706 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.123673 4706 scope.go:117] "RemoveContainer" containerID="d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.145411 4706 scope.go:117] "RemoveContainer" containerID="26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.171679 4706 scope.go:117] "RemoveContainer" containerID="bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.198395 4706 scope.go:117] "RemoveContainer" containerID="914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad" Dec 06 13:59:56 crc kubenswrapper[4706]: E1206 13:59:56.198864 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\": container with ID starting with 914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad not found: ID does not exist" containerID="914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.198930 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad"} err="failed to get container status \"914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\": rpc error: code = NotFound desc = could not find container \"914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad\": container with ID starting with 914c474de879212fdb769fcc8b572eccb24a440c3d3bbf4183d0bceaa9c89aad not found: ID does not exist" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.198966 4706 scope.go:117] "RemoveContainer" containerID="832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42" Dec 06 13:59:56 crc kubenswrapper[4706]: E1206 13:59:56.199562 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\": container with ID starting with 832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42 not found: ID does not exist" containerID="832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.199609 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42"} err="failed to get container status \"832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\": rpc error: code = NotFound desc = could not find container \"832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42\": container with ID starting with 832d1865154a1a01141d6ef3f96729c168d594fd46be7933a73ced9d43e5cb42 not found: ID does not exist" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.199637 4706 scope.go:117] "RemoveContainer" containerID="603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05" Dec 06 13:59:56 crc kubenswrapper[4706]: E1206 13:59:56.200165 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\": container with ID starting with 603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05 not found: ID does not exist" containerID="603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.200223 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05"} err="failed to get container status \"603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\": rpc error: code = NotFound desc = could not find container \"603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05\": container with ID starting with 603292d08f306ec42797981160134eb769b1e9481c76b75d6eabf1e6efb6df05 not found: ID does not exist" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.200265 4706 scope.go:117] "RemoveContainer" containerID="d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0" Dec 06 13:59:56 crc kubenswrapper[4706]: E1206 13:59:56.200797 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\": container with ID starting with d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0 not found: ID does not exist" containerID="d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.200843 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0"} err="failed to get container status \"d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\": rpc error: code = NotFound desc = could not find container \"d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0\": container with ID starting with d679eacf216dff31f38ed0fe85ad96906815dc799f0f8fdbe5f74376246d2ba0 not found: ID does not exist" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.200872 4706 scope.go:117] "RemoveContainer" containerID="26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788" Dec 06 13:59:56 crc kubenswrapper[4706]: E1206 13:59:56.201222 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\": container with ID starting with 26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788 not found: ID does not exist" containerID="26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.201281 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788"} err="failed to get container status \"26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\": rpc error: code = NotFound desc = could not find container \"26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788\": container with ID starting with 26f118608e0aac6c0b92fe2c7d3c5b47585fa20b6d6e493ac479d2a5002cf788 not found: ID does not exist" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.201370 4706 scope.go:117] "RemoveContainer" containerID="bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01" Dec 06 13:59:56 crc kubenswrapper[4706]: E1206 13:59:56.201742 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\": container with ID starting with bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01 not found: ID does not exist" containerID="bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01" Dec 06 13:59:56 crc kubenswrapper[4706]: I1206 13:59:56.201785 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01"} err="failed to get container status \"bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\": rpc error: code = NotFound desc = could not find container \"bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01\": container with ID starting with bbab6e701c1e5673ec6190b218c18fb13cd583e52fb9a25ab4f908d482257f01 not found: ID does not exist" Dec 06 13:59:57 crc kubenswrapper[4706]: I1206 13:59:57.513428 4706 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:57 crc kubenswrapper[4706]: I1206 13:59:57.515154 4706 status_manager.go:851] "Failed to get status for pod" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:57 crc kubenswrapper[4706]: I1206 13:59:57.523844 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 06 13:59:58 crc kubenswrapper[4706]: I1206 13:59:58.805541 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hjkcq" Dec 06 13:59:58 crc kubenswrapper[4706]: I1206 13:59:58.806486 4706 status_manager.go:851] "Failed to get status for pod" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" pod="openshift-marketplace/community-operators-hjkcq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hjkcq\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:58 crc kubenswrapper[4706]: I1206 13:59:58.807822 4706 status_manager.go:851] "Failed to get status for pod" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:59 crc kubenswrapper[4706]: I1206 13:59:59.154777 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zhlvc" Dec 06 13:59:59 crc kubenswrapper[4706]: I1206 13:59:59.155531 4706 status_manager.go:851] "Failed to get status for pod" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" pod="openshift-marketplace/community-operators-zhlvc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zhlvc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:59 crc kubenswrapper[4706]: I1206 13:59:59.155861 4706 status_manager.go:851] "Failed to get status for pod" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" pod="openshift-marketplace/community-operators-hjkcq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hjkcq\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:59 crc kubenswrapper[4706]: I1206 13:59:59.156131 4706 status_manager.go:851] "Failed to get status for pod" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:59 crc kubenswrapper[4706]: I1206 13:59:59.192856 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zhlvc" Dec 06 13:59:59 crc kubenswrapper[4706]: I1206 13:59:59.195032 4706 status_manager.go:851] "Failed to get status for pod" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:59 crc kubenswrapper[4706]: I1206 13:59:59.195464 4706 status_manager.go:851] "Failed to get status for pod" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" pod="openshift-marketplace/community-operators-zhlvc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zhlvc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 13:59:59 crc kubenswrapper[4706]: I1206 13:59:59.195812 4706 status_manager.go:851] "Failed to get status for pod" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" pod="openshift-marketplace/community-operators-hjkcq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hjkcq\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:00 crc kubenswrapper[4706]: E1206 14:00:00.212525 4706 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:00 crc kubenswrapper[4706]: E1206 14:00:00.213008 4706 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:00 crc kubenswrapper[4706]: E1206 14:00:00.213565 4706 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:00 crc kubenswrapper[4706]: E1206 14:00:00.213993 4706 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:00 crc kubenswrapper[4706]: E1206 14:00:00.214494 4706 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:00 crc kubenswrapper[4706]: I1206 14:00:00.214561 4706 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 06 14:00:00 crc kubenswrapper[4706]: E1206 14:00:00.215027 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="200ms" Dec 06 14:00:00 crc kubenswrapper[4706]: E1206 14:00:00.416137 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="400ms" Dec 06 14:00:00 crc kubenswrapper[4706]: E1206 14:00:00.818164 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="800ms" Dec 06 14:00:00 crc kubenswrapper[4706]: E1206 14:00:00.824787 4706 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.83:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187ea5106bdd0149 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 13:59:53.923518793 +0000 UTC m=+236.759311690,LastTimestamp:2025-12-06 13:59:53.923518793 +0000 UTC m=+236.759311690,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 14:00:01 crc kubenswrapper[4706]: E1206 14:00:01.176945 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T14:00:01Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T14:00:01Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T14:00:01Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T14:00:01Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:01 crc kubenswrapper[4706]: E1206 14:00:01.177551 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:01 crc kubenswrapper[4706]: E1206 14:00:01.177828 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:01 crc kubenswrapper[4706]: E1206 14:00:01.178247 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:01 crc kubenswrapper[4706]: E1206 14:00:01.178742 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:01 crc kubenswrapper[4706]: E1206 14:00:01.178770 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 14:00:01 crc kubenswrapper[4706]: E1206 14:00:01.619233 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="1.6s" Dec 06 14:00:01 crc kubenswrapper[4706]: I1206 14:00:01.963582 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x5pm6" Dec 06 14:00:01 crc kubenswrapper[4706]: I1206 14:00:01.964198 4706 status_manager.go:851] "Failed to get status for pod" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" pod="openshift-marketplace/community-operators-zhlvc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zhlvc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:01 crc kubenswrapper[4706]: I1206 14:00:01.964967 4706 status_manager.go:851] "Failed to get status for pod" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" pod="openshift-marketplace/community-operators-hjkcq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hjkcq\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:01 crc kubenswrapper[4706]: I1206 14:00:01.965402 4706 status_manager.go:851] "Failed to get status for pod" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" pod="openshift-marketplace/redhat-operators-x5pm6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x5pm6\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:01 crc kubenswrapper[4706]: I1206 14:00:01.965867 4706 status_manager.go:851] "Failed to get status for pod" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:02 crc kubenswrapper[4706]: I1206 14:00:02.013863 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x5pm6" Dec 06 14:00:02 crc kubenswrapper[4706]: I1206 14:00:02.014726 4706 status_manager.go:851] "Failed to get status for pod" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" pod="openshift-marketplace/community-operators-zhlvc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zhlvc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:02 crc kubenswrapper[4706]: I1206 14:00:02.015258 4706 status_manager.go:851] "Failed to get status for pod" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" pod="openshift-marketplace/community-operators-hjkcq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hjkcq\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:02 crc kubenswrapper[4706]: I1206 14:00:02.016227 4706 status_manager.go:851] "Failed to get status for pod" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" pod="openshift-marketplace/redhat-operators-x5pm6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x5pm6\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:02 crc kubenswrapper[4706]: I1206 14:00:02.016790 4706 status_manager.go:851] "Failed to get status for pod" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:03 crc kubenswrapper[4706]: E1206 14:00:03.221023 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="3.2s" Dec 06 14:00:06 crc kubenswrapper[4706]: E1206 14:00:06.422269 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.83:6443: connect: connection refused" interval="6.4s" Dec 06 14:00:07 crc kubenswrapper[4706]: I1206 14:00:07.516566 4706 status_manager.go:851] "Failed to get status for pod" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" pod="openshift-marketplace/community-operators-zhlvc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zhlvc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:07 crc kubenswrapper[4706]: I1206 14:00:07.518673 4706 status_manager.go:851] "Failed to get status for pod" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" pod="openshift-marketplace/community-operators-hjkcq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hjkcq\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:07 crc kubenswrapper[4706]: I1206 14:00:07.519361 4706 status_manager.go:851] "Failed to get status for pod" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" pod="openshift-marketplace/redhat-operators-x5pm6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x5pm6\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:07 crc kubenswrapper[4706]: I1206 14:00:07.519878 4706 status_manager.go:851] "Failed to get status for pod" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.135215 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.135348 4706 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5" exitCode=1 Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.135394 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5"} Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.136278 4706 scope.go:117] "RemoveContainer" containerID="9a8235a3a34e707fd50261f7ac9613ae463f29698d40a03298338fe5bd830ed5" Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.136653 4706 status_manager.go:851] "Failed to get status for pod" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" pod="openshift-marketplace/community-operators-zhlvc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zhlvc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.137290 4706 status_manager.go:851] "Failed to get status for pod" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" pod="openshift-marketplace/community-operators-hjkcq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hjkcq\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.137976 4706 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.138521 4706 status_manager.go:851] "Failed to get status for pod" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" pod="openshift-marketplace/redhat-operators-x5pm6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x5pm6\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.139104 4706 status_manager.go:851] "Failed to get status for pod" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.510175 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.511270 4706 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.511933 4706 status_manager.go:851] "Failed to get status for pod" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.512566 4706 status_manager.go:851] "Failed to get status for pod" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" pod="openshift-marketplace/redhat-operators-x5pm6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x5pm6\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.513099 4706 status_manager.go:851] "Failed to get status for pod" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" pod="openshift-marketplace/community-operators-zhlvc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zhlvc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.513589 4706 status_manager.go:851] "Failed to get status for pod" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" pod="openshift-marketplace/community-operators-hjkcq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hjkcq\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.525754 4706 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e865c24c-0cb5-40dc-8b60-d30a6e80780d" Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.525820 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e865c24c-0cb5-40dc-8b60-d30a6e80780d" Dec 06 14:00:08 crc kubenswrapper[4706]: E1206 14:00:08.526241 4706 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 14:00:08 crc kubenswrapper[4706]: I1206 14:00:08.527122 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 14:00:08 crc kubenswrapper[4706]: W1206 14:00:08.549647 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-6392af868153526c59930a06e247d54ec2d75af622b58d5128142600c96524b3 WatchSource:0}: Error finding container 6392af868153526c59930a06e247d54ec2d75af622b58d5128142600c96524b3: Status 404 returned error can't find the container with id 6392af868153526c59930a06e247d54ec2d75af622b58d5128142600c96524b3 Dec 06 14:00:09 crc kubenswrapper[4706]: I1206 14:00:09.149873 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 14:00:09 crc kubenswrapper[4706]: I1206 14:00:09.150044 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"87eeba9ad30ed42cb9c31024ffd15dccd0c2f67684ecfe9cd1625d012805fec5"} Dec 06 14:00:09 crc kubenswrapper[4706]: I1206 14:00:09.151468 4706 status_manager.go:851] "Failed to get status for pod" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" pod="openshift-marketplace/community-operators-zhlvc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zhlvc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:09 crc kubenswrapper[4706]: I1206 14:00:09.152184 4706 status_manager.go:851] "Failed to get status for pod" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" pod="openshift-marketplace/community-operators-hjkcq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hjkcq\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:09 crc kubenswrapper[4706]: I1206 14:00:09.152805 4706 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:09 crc kubenswrapper[4706]: I1206 14:00:09.153301 4706 status_manager.go:851] "Failed to get status for pod" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" pod="openshift-marketplace/redhat-operators-x5pm6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x5pm6\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:09 crc kubenswrapper[4706]: I1206 14:00:09.153968 4706 status_manager.go:851] "Failed to get status for pod" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:09 crc kubenswrapper[4706]: I1206 14:00:09.155851 4706 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="0e1c11a2265aaa3fe336b40471e57d74462cdd960896f9b85eb795d219a48fcd" exitCode=0 Dec 06 14:00:09 crc kubenswrapper[4706]: I1206 14:00:09.155939 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"0e1c11a2265aaa3fe336b40471e57d74462cdd960896f9b85eb795d219a48fcd"} Dec 06 14:00:09 crc kubenswrapper[4706]: I1206 14:00:09.156015 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6392af868153526c59930a06e247d54ec2d75af622b58d5128142600c96524b3"} Dec 06 14:00:09 crc kubenswrapper[4706]: I1206 14:00:09.156616 4706 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e865c24c-0cb5-40dc-8b60-d30a6e80780d" Dec 06 14:00:09 crc kubenswrapper[4706]: I1206 14:00:09.156662 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e865c24c-0cb5-40dc-8b60-d30a6e80780d" Dec 06 14:00:09 crc kubenswrapper[4706]: I1206 14:00:09.157267 4706 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:09 crc kubenswrapper[4706]: E1206 14:00:09.157471 4706 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 14:00:09 crc kubenswrapper[4706]: I1206 14:00:09.157878 4706 status_manager.go:851] "Failed to get status for pod" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:09 crc kubenswrapper[4706]: I1206 14:00:09.158745 4706 status_manager.go:851] "Failed to get status for pod" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" pod="openshift-marketplace/redhat-operators-x5pm6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x5pm6\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:09 crc kubenswrapper[4706]: I1206 14:00:09.159364 4706 status_manager.go:851] "Failed to get status for pod" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" pod="openshift-marketplace/community-operators-zhlvc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zhlvc\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:09 crc kubenswrapper[4706]: I1206 14:00:09.160020 4706 status_manager.go:851] "Failed to get status for pod" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" pod="openshift-marketplace/community-operators-hjkcq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hjkcq\": dial tcp 38.102.83.83:6443: connect: connection refused" Dec 06 14:00:10 crc kubenswrapper[4706]: I1206 14:00:10.169807 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f1d4cd9fac9e022f053cbab93256a4c7758677a987ac3414ad07855f5c658228"} Dec 06 14:00:10 crc kubenswrapper[4706]: I1206 14:00:10.171036 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4216c60d5da57f131b98a5384fa1474a9d7f93723fa328275e5c02ce76ba0a6e"} Dec 06 14:00:10 crc kubenswrapper[4706]: I1206 14:00:10.171062 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"18f40c5177df3413727ec4af21d41c5a22d128e3ae394bf334d66da58029ae6a"} Dec 06 14:00:11 crc kubenswrapper[4706]: I1206 14:00:11.185748 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"16b3e9c316c4e1d4a477645e75281cea2406a4255bae365a81eb440d5a67b1c0"} Dec 06 14:00:11 crc kubenswrapper[4706]: I1206 14:00:11.185801 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f8a5c005fcf02e5af0cb6df3d4dc8cea01b90a4fda23e82592d36e69135f6ed6"} Dec 06 14:00:11 crc kubenswrapper[4706]: I1206 14:00:11.186001 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 14:00:11 crc kubenswrapper[4706]: I1206 14:00:11.186237 4706 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e865c24c-0cb5-40dc-8b60-d30a6e80780d" Dec 06 14:00:11 crc kubenswrapper[4706]: I1206 14:00:11.186285 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e865c24c-0cb5-40dc-8b60-d30a6e80780d" Dec 06 14:00:11 crc kubenswrapper[4706]: I1206 14:00:11.945246 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 14:00:13 crc kubenswrapper[4706]: I1206 14:00:13.527840 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 14:00:13 crc kubenswrapper[4706]: I1206 14:00:13.528462 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 14:00:13 crc kubenswrapper[4706]: I1206 14:00:13.536624 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 14:00:14 crc kubenswrapper[4706]: I1206 14:00:14.391059 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 14:00:14 crc kubenswrapper[4706]: I1206 14:00:14.394375 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 14:00:16 crc kubenswrapper[4706]: I1206 14:00:16.326495 4706 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 14:00:17 crc kubenswrapper[4706]: I1206 14:00:17.223476 4706 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e865c24c-0cb5-40dc-8b60-d30a6e80780d" Dec 06 14:00:17 crc kubenswrapper[4706]: I1206 14:00:17.223509 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e865c24c-0cb5-40dc-8b60-d30a6e80780d" Dec 06 14:00:17 crc kubenswrapper[4706]: I1206 14:00:17.228841 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 14:00:17 crc kubenswrapper[4706]: I1206 14:00:17.546721 4706 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="992058f3-ffd0-402e-a748-e220500601c4" Dec 06 14:00:18 crc kubenswrapper[4706]: I1206 14:00:18.229878 4706 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e865c24c-0cb5-40dc-8b60-d30a6e80780d" Dec 06 14:00:18 crc kubenswrapper[4706]: I1206 14:00:18.229908 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e865c24c-0cb5-40dc-8b60-d30a6e80780d" Dec 06 14:00:18 crc kubenswrapper[4706]: I1206 14:00:18.233082 4706 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="992058f3-ffd0-402e-a748-e220500601c4" Dec 06 14:00:21 crc kubenswrapper[4706]: I1206 14:00:21.952913 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 14:00:26 crc kubenswrapper[4706]: I1206 14:00:26.035636 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 14:00:26 crc kubenswrapper[4706]: I1206 14:00:26.138400 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 14:00:26 crc kubenswrapper[4706]: I1206 14:00:26.248625 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 14:00:26 crc kubenswrapper[4706]: I1206 14:00:26.378056 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 14:00:26 crc kubenswrapper[4706]: I1206 14:00:26.758446 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 14:00:26 crc kubenswrapper[4706]: I1206 14:00:26.804977 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 14:00:26 crc kubenswrapper[4706]: I1206 14:00:26.927091 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 14:00:26 crc kubenswrapper[4706]: I1206 14:00:26.974476 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 14:00:26 crc kubenswrapper[4706]: I1206 14:00:26.974934 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 14:00:27 crc kubenswrapper[4706]: I1206 14:00:27.053736 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 14:00:27 crc kubenswrapper[4706]: I1206 14:00:27.404624 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 14:00:27 crc kubenswrapper[4706]: I1206 14:00:27.447041 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 14:00:27 crc kubenswrapper[4706]: I1206 14:00:27.555386 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 14:00:27 crc kubenswrapper[4706]: I1206 14:00:27.594293 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 14:00:27 crc kubenswrapper[4706]: I1206 14:00:27.659881 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 14:00:27 crc kubenswrapper[4706]: I1206 14:00:27.668183 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 14:00:27 crc kubenswrapper[4706]: I1206 14:00:27.872256 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 14:00:28 crc kubenswrapper[4706]: I1206 14:00:28.082882 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 14:00:28 crc kubenswrapper[4706]: I1206 14:00:28.371504 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 14:00:28 crc kubenswrapper[4706]: I1206 14:00:28.392546 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 14:00:28 crc kubenswrapper[4706]: I1206 14:00:28.433058 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 14:00:28 crc kubenswrapper[4706]: I1206 14:00:28.450400 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 14:00:28 crc kubenswrapper[4706]: I1206 14:00:28.582716 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 14:00:28 crc kubenswrapper[4706]: I1206 14:00:28.858299 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 14:00:28 crc kubenswrapper[4706]: I1206 14:00:28.981604 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 14:00:29 crc kubenswrapper[4706]: I1206 14:00:29.152863 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 14:00:29 crc kubenswrapper[4706]: I1206 14:00:29.228144 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 14:00:29 crc kubenswrapper[4706]: I1206 14:00:29.358509 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 14:00:29 crc kubenswrapper[4706]: I1206 14:00:29.361648 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 14:00:29 crc kubenswrapper[4706]: I1206 14:00:29.391263 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 14:00:29 crc kubenswrapper[4706]: I1206 14:00:29.471755 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 14:00:29 crc kubenswrapper[4706]: I1206 14:00:29.552534 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 14:00:29 crc kubenswrapper[4706]: I1206 14:00:29.562024 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 14:00:29 crc kubenswrapper[4706]: I1206 14:00:29.562401 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 14:00:29 crc kubenswrapper[4706]: I1206 14:00:29.562795 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 14:00:29 crc kubenswrapper[4706]: I1206 14:00:29.711925 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 14:00:29 crc kubenswrapper[4706]: I1206 14:00:29.762491 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 14:00:29 crc kubenswrapper[4706]: I1206 14:00:29.769226 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 14:00:29 crc kubenswrapper[4706]: I1206 14:00:29.878073 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 14:00:29 crc kubenswrapper[4706]: I1206 14:00:29.898806 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 14:00:29 crc kubenswrapper[4706]: I1206 14:00:29.909529 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 14:00:30 crc kubenswrapper[4706]: I1206 14:00:30.045678 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 14:00:30 crc kubenswrapper[4706]: I1206 14:00:30.124025 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 14:00:30 crc kubenswrapper[4706]: I1206 14:00:30.197062 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 14:00:30 crc kubenswrapper[4706]: I1206 14:00:30.369732 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 14:00:30 crc kubenswrapper[4706]: I1206 14:00:30.386131 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 14:00:30 crc kubenswrapper[4706]: I1206 14:00:30.419740 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 14:00:30 crc kubenswrapper[4706]: I1206 14:00:30.421420 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 14:00:30 crc kubenswrapper[4706]: I1206 14:00:30.427787 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 14:00:30 crc kubenswrapper[4706]: I1206 14:00:30.432079 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 14:00:30 crc kubenswrapper[4706]: I1206 14:00:30.465472 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 14:00:30 crc kubenswrapper[4706]: I1206 14:00:30.479180 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 14:00:30 crc kubenswrapper[4706]: I1206 14:00:30.546673 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 14:00:30 crc kubenswrapper[4706]: I1206 14:00:30.613553 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 14:00:30 crc kubenswrapper[4706]: I1206 14:00:30.798267 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 14:00:30 crc kubenswrapper[4706]: I1206 14:00:30.961273 4706 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 14:00:31 crc kubenswrapper[4706]: I1206 14:00:31.067212 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 14:00:31 crc kubenswrapper[4706]: I1206 14:00:31.068744 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 14:00:31 crc kubenswrapper[4706]: I1206 14:00:31.126501 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 14:00:31 crc kubenswrapper[4706]: I1206 14:00:31.180730 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 14:00:31 crc kubenswrapper[4706]: I1206 14:00:31.278744 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 14:00:31 crc kubenswrapper[4706]: I1206 14:00:31.369084 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 14:00:31 crc kubenswrapper[4706]: I1206 14:00:31.458820 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 14:00:31 crc kubenswrapper[4706]: I1206 14:00:31.544566 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 14:00:31 crc kubenswrapper[4706]: I1206 14:00:31.553710 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 14:00:31 crc kubenswrapper[4706]: I1206 14:00:31.700194 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 14:00:31 crc kubenswrapper[4706]: I1206 14:00:31.801205 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 14:00:31 crc kubenswrapper[4706]: I1206 14:00:31.832482 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.003535 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.078111 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.103517 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.109729 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.216464 4706 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.224287 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.224487 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.231104 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.232084 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.260402 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.260373081 podStartE2EDuration="16.260373081s" podCreationTimestamp="2025-12-06 14:00:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:00:32.255028758 +0000 UTC m=+275.090821655" watchObservedRunningTime="2025-12-06 14:00:32.260373081 +0000 UTC m=+275.096165968" Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.269095 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.271297 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.311901 4706 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.311943 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.400791 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.532803 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.595826 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.681237 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.700707 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 14:00:32 crc kubenswrapper[4706]: I1206 14:00:32.884913 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.104364 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.177874 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.382089 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.400581 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.404414 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.451740 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.471182 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.506832 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.553358 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.557649 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.584953 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt"] Dec 06 14:00:33 crc kubenswrapper[4706]: E1206 14:00:33.585220 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" containerName="installer" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.585238 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" containerName="installer" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.585377 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="40d27479-a513-43d4-8ad7-9a4daa8a29d7" containerName="installer" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.585837 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.586380 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.588115 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.588468 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.599105 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt"] Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.685093 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.727529 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.728012 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.771756 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzmg7\" (UniqueName: \"kubernetes.io/projected/1a678c56-c771-4368-9712-68dd02c279dd-kube-api-access-bzmg7\") pod \"collect-profiles-29417160-596lt\" (UID: \"1a678c56-c771-4368-9712-68dd02c279dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.771813 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a678c56-c771-4368-9712-68dd02c279dd-secret-volume\") pod \"collect-profiles-29417160-596lt\" (UID: \"1a678c56-c771-4368-9712-68dd02c279dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.771846 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a678c56-c771-4368-9712-68dd02c279dd-config-volume\") pod \"collect-profiles-29417160-596lt\" (UID: \"1a678c56-c771-4368-9712-68dd02c279dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.775066 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.811165 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.832718 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.887457 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzmg7\" (UniqueName: \"kubernetes.io/projected/1a678c56-c771-4368-9712-68dd02c279dd-kube-api-access-bzmg7\") pod \"collect-profiles-29417160-596lt\" (UID: \"1a678c56-c771-4368-9712-68dd02c279dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.887523 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a678c56-c771-4368-9712-68dd02c279dd-secret-volume\") pod \"collect-profiles-29417160-596lt\" (UID: \"1a678c56-c771-4368-9712-68dd02c279dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.887578 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a678c56-c771-4368-9712-68dd02c279dd-config-volume\") pod \"collect-profiles-29417160-596lt\" (UID: \"1a678c56-c771-4368-9712-68dd02c279dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.888810 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a678c56-c771-4368-9712-68dd02c279dd-config-volume\") pod \"collect-profiles-29417160-596lt\" (UID: \"1a678c56-c771-4368-9712-68dd02c279dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.890043 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.904532 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a678c56-c771-4368-9712-68dd02c279dd-secret-volume\") pod \"collect-profiles-29417160-596lt\" (UID: \"1a678c56-c771-4368-9712-68dd02c279dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.911959 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzmg7\" (UniqueName: \"kubernetes.io/projected/1a678c56-c771-4368-9712-68dd02c279dd-kube-api-access-bzmg7\") pod \"collect-profiles-29417160-596lt\" (UID: \"1a678c56-c771-4368-9712-68dd02c279dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt" Dec 06 14:00:33 crc kubenswrapper[4706]: I1206 14:00:33.952032 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.016142 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.053664 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.065228 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.130292 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.134637 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.201605 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.258007 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.259843 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.311347 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.395734 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.454340 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.482196 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.637899 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.662461 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.679243 4706 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.681223 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt"] Dec 06 14:00:34 crc kubenswrapper[4706]: W1206 14:00:34.691680 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a678c56_c771_4368_9712_68dd02c279dd.slice/crio-3e134840bc2aa4500b0eb362598b9e5fddbd663583db6b582df4f00cb5b31e85 WatchSource:0}: Error finding container 3e134840bc2aa4500b0eb362598b9e5fddbd663583db6b582df4f00cb5b31e85: Status 404 returned error can't find the container with id 3e134840bc2aa4500b0eb362598b9e5fddbd663583db6b582df4f00cb5b31e85 Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.764932 4706 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.775702 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.785799 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.942246 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.955950 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.962754 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 14:00:34 crc kubenswrapper[4706]: I1206 14:00:34.979803 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.022791 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.092960 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.093253 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.155451 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.194380 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.200011 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.226824 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.351602 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_collect-profiles-29417160-596lt_1a678c56-c771-4368-9712-68dd02c279dd/collect-profiles/0.log" Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.352063 4706 generic.go:334] "Generic (PLEG): container finished" podID="1a678c56-c771-4368-9712-68dd02c279dd" containerID="7a423baf74a6dbbe4e6a9509a6e6f3e323257a678c85ea5bf4c6ffa5610853c7" exitCode=1 Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.352114 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt" event={"ID":"1a678c56-c771-4368-9712-68dd02c279dd","Type":"ContainerDied","Data":"7a423baf74a6dbbe4e6a9509a6e6f3e323257a678c85ea5bf4c6ffa5610853c7"} Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.352488 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt" event={"ID":"1a678c56-c771-4368-9712-68dd02c279dd","Type":"ContainerStarted","Data":"3e134840bc2aa4500b0eb362598b9e5fddbd663583db6b582df4f00cb5b31e85"} Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.467768 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.583067 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.586078 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.711694 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.721080 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.733127 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.750871 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.775882 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 14:00:35 crc kubenswrapper[4706]: I1206 14:00:35.864540 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.002869 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.025248 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.034071 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.041259 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.126201 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.146662 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.153049 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.284282 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.289939 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.300784 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.348573 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.415892 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.416014 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.429974 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.433875 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.485730 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.546115 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.596792 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.684682 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_collect-profiles-29417160-596lt_1a678c56-c771-4368-9712-68dd02c279dd/collect-profiles/0.log" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.684793 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.720801 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.769628 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.774481 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.828640 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a678c56-c771-4368-9712-68dd02c279dd-secret-volume\") pod \"1a678c56-c771-4368-9712-68dd02c279dd\" (UID: \"1a678c56-c771-4368-9712-68dd02c279dd\") " Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.828686 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a678c56-c771-4368-9712-68dd02c279dd-config-volume\") pod \"1a678c56-c771-4368-9712-68dd02c279dd\" (UID: \"1a678c56-c771-4368-9712-68dd02c279dd\") " Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.828752 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzmg7\" (UniqueName: \"kubernetes.io/projected/1a678c56-c771-4368-9712-68dd02c279dd-kube-api-access-bzmg7\") pod \"1a678c56-c771-4368-9712-68dd02c279dd\" (UID: \"1a678c56-c771-4368-9712-68dd02c279dd\") " Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.830051 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a678c56-c771-4368-9712-68dd02c279dd-config-volume" (OuterVolumeSpecName: "config-volume") pod "1a678c56-c771-4368-9712-68dd02c279dd" (UID: "1a678c56-c771-4368-9712-68dd02c279dd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.835409 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a678c56-c771-4368-9712-68dd02c279dd-kube-api-access-bzmg7" (OuterVolumeSpecName: "kube-api-access-bzmg7") pod "1a678c56-c771-4368-9712-68dd02c279dd" (UID: "1a678c56-c771-4368-9712-68dd02c279dd"). InnerVolumeSpecName "kube-api-access-bzmg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.836607 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a678c56-c771-4368-9712-68dd02c279dd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1a678c56-c771-4368-9712-68dd02c279dd" (UID: "1a678c56-c771-4368-9712-68dd02c279dd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.845488 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.856984 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.863567 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.929735 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.929856 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzmg7\" (UniqueName: \"kubernetes.io/projected/1a678c56-c771-4368-9712-68dd02c279dd-kube-api-access-bzmg7\") on node \"crc\" DevicePath \"\"" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.929924 4706 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a678c56-c771-4368-9712-68dd02c279dd-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 14:00:36 crc kubenswrapper[4706]: I1206 14:00:36.929940 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a678c56-c771-4368-9712-68dd02c279dd-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.048167 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.085369 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.100303 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.112531 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.121207 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.145176 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.262028 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.355424 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.362648 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.370757 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_collect-profiles-29417160-596lt_1a678c56-c771-4368-9712-68dd02c279dd/collect-profiles/0.log" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.370811 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt" event={"ID":"1a678c56-c771-4368-9712-68dd02c279dd","Type":"ContainerDied","Data":"3e134840bc2aa4500b0eb362598b9e5fddbd663583db6b582df4f00cb5b31e85"} Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.370835 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e134840bc2aa4500b0eb362598b9e5fddbd663583db6b582df4f00cb5b31e85" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.370883 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-596lt" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.393051 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.441498 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.466123 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.504628 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.510839 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.538179 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.540253 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.636889 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.639123 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.670852 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.691268 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.692462 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.769296 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.783852 4706 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.784189 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://1aa02db026a29f6ba7f04e7271a176bb36d434dd04cefb0511151b5520f6579b" gracePeriod=5 Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.871016 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.882919 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.895504 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 14:00:37 crc kubenswrapper[4706]: I1206 14:00:37.962278 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.065284 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.125054 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.138228 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.178814 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.183527 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.220924 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.236205 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.252597 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.267540 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.272990 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.534886 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.535248 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.607521 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.630773 4706 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.673749 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.682109 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.705530 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.720920 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.800485 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.888991 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 14:00:38 crc kubenswrapper[4706]: I1206 14:00:38.967698 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 14:00:39 crc kubenswrapper[4706]: I1206 14:00:39.066155 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 14:00:39 crc kubenswrapper[4706]: I1206 14:00:39.067605 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 14:00:39 crc kubenswrapper[4706]: I1206 14:00:39.113096 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 14:00:39 crc kubenswrapper[4706]: I1206 14:00:39.149028 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 14:00:39 crc kubenswrapper[4706]: I1206 14:00:39.262246 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 14:00:39 crc kubenswrapper[4706]: I1206 14:00:39.335016 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 14:00:39 crc kubenswrapper[4706]: I1206 14:00:39.509599 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 14:00:39 crc kubenswrapper[4706]: I1206 14:00:39.649114 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 14:00:39 crc kubenswrapper[4706]: I1206 14:00:39.691606 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 14:00:39 crc kubenswrapper[4706]: I1206 14:00:39.691666 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 14:00:39 crc kubenswrapper[4706]: I1206 14:00:39.712625 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 14:00:39 crc kubenswrapper[4706]: I1206 14:00:39.813579 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 14:00:39 crc kubenswrapper[4706]: I1206 14:00:39.903841 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 14:00:39 crc kubenswrapper[4706]: I1206 14:00:39.993471 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 14:00:40 crc kubenswrapper[4706]: I1206 14:00:40.013584 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 14:00:40 crc kubenswrapper[4706]: I1206 14:00:40.037361 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 14:00:40 crc kubenswrapper[4706]: I1206 14:00:40.042297 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 14:00:40 crc kubenswrapper[4706]: I1206 14:00:40.211367 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 14:00:40 crc kubenswrapper[4706]: I1206 14:00:40.287117 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 14:00:40 crc kubenswrapper[4706]: I1206 14:00:40.348624 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 14:00:40 crc kubenswrapper[4706]: I1206 14:00:40.427623 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 14:00:40 crc kubenswrapper[4706]: I1206 14:00:40.541747 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 14:00:40 crc kubenswrapper[4706]: I1206 14:00:40.567726 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 14:00:40 crc kubenswrapper[4706]: I1206 14:00:40.690560 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 14:00:40 crc kubenswrapper[4706]: I1206 14:00:40.721129 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 14:00:40 crc kubenswrapper[4706]: I1206 14:00:40.772127 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 14:00:40 crc kubenswrapper[4706]: I1206 14:00:40.824487 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 14:00:40 crc kubenswrapper[4706]: I1206 14:00:40.872627 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 14:00:40 crc kubenswrapper[4706]: I1206 14:00:40.876503 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 14:00:41 crc kubenswrapper[4706]: I1206 14:00:41.001131 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 14:00:41 crc kubenswrapper[4706]: I1206 14:00:41.369948 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 14:00:41 crc kubenswrapper[4706]: I1206 14:00:41.373022 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 14:00:41 crc kubenswrapper[4706]: I1206 14:00:41.647737 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 14:00:41 crc kubenswrapper[4706]: I1206 14:00:41.993430 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 14:00:42 crc kubenswrapper[4706]: I1206 14:00:42.578324 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.371843 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.371952 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.414591 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.414650 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.414706 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.414765 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.414798 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.415172 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.415163 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.415228 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.415249 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.415602 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.415689 4706 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="1aa02db026a29f6ba7f04e7271a176bb36d434dd04cefb0511151b5520f6579b" exitCode=137 Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.415753 4706 scope.go:117] "RemoveContainer" containerID="1aa02db026a29f6ba7f04e7271a176bb36d434dd04cefb0511151b5520f6579b" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.415957 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.425331 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.463959 4706 scope.go:117] "RemoveContainer" containerID="1aa02db026a29f6ba7f04e7271a176bb36d434dd04cefb0511151b5520f6579b" Dec 06 14:00:43 crc kubenswrapper[4706]: E1206 14:00:43.464530 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1aa02db026a29f6ba7f04e7271a176bb36d434dd04cefb0511151b5520f6579b\": container with ID starting with 1aa02db026a29f6ba7f04e7271a176bb36d434dd04cefb0511151b5520f6579b not found: ID does not exist" containerID="1aa02db026a29f6ba7f04e7271a176bb36d434dd04cefb0511151b5520f6579b" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.464561 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1aa02db026a29f6ba7f04e7271a176bb36d434dd04cefb0511151b5520f6579b"} err="failed to get container status \"1aa02db026a29f6ba7f04e7271a176bb36d434dd04cefb0511151b5520f6579b\": rpc error: code = NotFound desc = could not find container \"1aa02db026a29f6ba7f04e7271a176bb36d434dd04cefb0511151b5520f6579b\": container with ID starting with 1aa02db026a29f6ba7f04e7271a176bb36d434dd04cefb0511151b5520f6579b not found: ID does not exist" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.516755 4706 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.516847 4706 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.516869 4706 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.516886 4706 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.516944 4706 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 14:00:43 crc kubenswrapper[4706]: I1206 14:00:43.524586 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.029252 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw"] Dec 06 14:00:44 crc kubenswrapper[4706]: E1206 14:00:44.031772 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.031818 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 14:00:44 crc kubenswrapper[4706]: E1206 14:00:44.031856 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a678c56-c771-4368-9712-68dd02c279dd" containerName="collect-profiles" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.031873 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a678c56-c771-4368-9712-68dd02c279dd" containerName="collect-profiles" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.032051 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a678c56-c771-4368-9712-68dd02c279dd" containerName="collect-profiles" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.032078 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.032847 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.035508 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw"] Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.037076 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.037348 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.125443 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4fft\" (UniqueName: \"kubernetes.io/projected/9944885a-bc84-4efb-b0b5-6d850a4df958-kube-api-access-k4fft\") pod \"collect-profiles-29417160-mfxrw\" (UID: \"9944885a-bc84-4efb-b0b5-6d850a4df958\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.126086 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9944885a-bc84-4efb-b0b5-6d850a4df958-config-volume\") pod \"collect-profiles-29417160-mfxrw\" (UID: \"9944885a-bc84-4efb-b0b5-6d850a4df958\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.126140 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9944885a-bc84-4efb-b0b5-6d850a4df958-secret-volume\") pod \"collect-profiles-29417160-mfxrw\" (UID: \"9944885a-bc84-4efb-b0b5-6d850a4df958\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.227477 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4fft\" (UniqueName: \"kubernetes.io/projected/9944885a-bc84-4efb-b0b5-6d850a4df958-kube-api-access-k4fft\") pod \"collect-profiles-29417160-mfxrw\" (UID: \"9944885a-bc84-4efb-b0b5-6d850a4df958\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.227577 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9944885a-bc84-4efb-b0b5-6d850a4df958-config-volume\") pod \"collect-profiles-29417160-mfxrw\" (UID: \"9944885a-bc84-4efb-b0b5-6d850a4df958\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.227640 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9944885a-bc84-4efb-b0b5-6d850a4df958-secret-volume\") pod \"collect-profiles-29417160-mfxrw\" (UID: \"9944885a-bc84-4efb-b0b5-6d850a4df958\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.229433 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9944885a-bc84-4efb-b0b5-6d850a4df958-config-volume\") pod \"collect-profiles-29417160-mfxrw\" (UID: \"9944885a-bc84-4efb-b0b5-6d850a4df958\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.234244 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9944885a-bc84-4efb-b0b5-6d850a4df958-secret-volume\") pod \"collect-profiles-29417160-mfxrw\" (UID: \"9944885a-bc84-4efb-b0b5-6d850a4df958\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.259020 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4fft\" (UniqueName: \"kubernetes.io/projected/9944885a-bc84-4efb-b0b5-6d850a4df958-kube-api-access-k4fft\") pod \"collect-profiles-29417160-mfxrw\" (UID: \"9944885a-bc84-4efb-b0b5-6d850a4df958\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.361228 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw" Dec 06 14:00:44 crc kubenswrapper[4706]: I1206 14:00:44.652833 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw"] Dec 06 14:00:45 crc kubenswrapper[4706]: I1206 14:00:45.437014 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_collect-profiles-29417160-mfxrw_9944885a-bc84-4efb-b0b5-6d850a4df958/collect-profiles/0.log" Dec 06 14:00:45 crc kubenswrapper[4706]: I1206 14:00:45.437782 4706 generic.go:334] "Generic (PLEG): container finished" podID="9944885a-bc84-4efb-b0b5-6d850a4df958" containerID="f875551708a29715c6e59b4f1cdfc505ae12b3ed05f59a29ec86fe76a255d09b" exitCode=1 Dec 06 14:00:45 crc kubenswrapper[4706]: I1206 14:00:45.437839 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw" event={"ID":"9944885a-bc84-4efb-b0b5-6d850a4df958","Type":"ContainerDied","Data":"f875551708a29715c6e59b4f1cdfc505ae12b3ed05f59a29ec86fe76a255d09b"} Dec 06 14:00:45 crc kubenswrapper[4706]: I1206 14:00:45.437890 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw" event={"ID":"9944885a-bc84-4efb-b0b5-6d850a4df958","Type":"ContainerStarted","Data":"0d188d8a26e5ce71fe14436dd9c32d03358062685b4ee85332d526a68e028a66"} Dec 06 14:00:46 crc kubenswrapper[4706]: I1206 14:00:46.769868 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_collect-profiles-29417160-mfxrw_9944885a-bc84-4efb-b0b5-6d850a4df958/collect-profiles/0.log" Dec 06 14:00:46 crc kubenswrapper[4706]: I1206 14:00:46.770392 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw" Dec 06 14:00:46 crc kubenswrapper[4706]: I1206 14:00:46.966263 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9944885a-bc84-4efb-b0b5-6d850a4df958-config-volume\") pod \"9944885a-bc84-4efb-b0b5-6d850a4df958\" (UID: \"9944885a-bc84-4efb-b0b5-6d850a4df958\") " Dec 06 14:00:46 crc kubenswrapper[4706]: I1206 14:00:46.966453 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9944885a-bc84-4efb-b0b5-6d850a4df958-secret-volume\") pod \"9944885a-bc84-4efb-b0b5-6d850a4df958\" (UID: \"9944885a-bc84-4efb-b0b5-6d850a4df958\") " Dec 06 14:00:46 crc kubenswrapper[4706]: I1206 14:00:46.966503 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4fft\" (UniqueName: \"kubernetes.io/projected/9944885a-bc84-4efb-b0b5-6d850a4df958-kube-api-access-k4fft\") pod \"9944885a-bc84-4efb-b0b5-6d850a4df958\" (UID: \"9944885a-bc84-4efb-b0b5-6d850a4df958\") " Dec 06 14:00:46 crc kubenswrapper[4706]: I1206 14:00:46.967511 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9944885a-bc84-4efb-b0b5-6d850a4df958-config-volume" (OuterVolumeSpecName: "config-volume") pod "9944885a-bc84-4efb-b0b5-6d850a4df958" (UID: "9944885a-bc84-4efb-b0b5-6d850a4df958"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:00:46 crc kubenswrapper[4706]: I1206 14:00:46.968093 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9944885a-bc84-4efb-b0b5-6d850a4df958-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 14:00:46 crc kubenswrapper[4706]: I1206 14:00:46.975517 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9944885a-bc84-4efb-b0b5-6d850a4df958-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9944885a-bc84-4efb-b0b5-6d850a4df958" (UID: "9944885a-bc84-4efb-b0b5-6d850a4df958"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:00:46 crc kubenswrapper[4706]: I1206 14:00:46.975643 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9944885a-bc84-4efb-b0b5-6d850a4df958-kube-api-access-k4fft" (OuterVolumeSpecName: "kube-api-access-k4fft") pod "9944885a-bc84-4efb-b0b5-6d850a4df958" (UID: "9944885a-bc84-4efb-b0b5-6d850a4df958"). InnerVolumeSpecName "kube-api-access-k4fft". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:00:47 crc kubenswrapper[4706]: I1206 14:00:47.068941 4706 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9944885a-bc84-4efb-b0b5-6d850a4df958-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 14:00:47 crc kubenswrapper[4706]: I1206 14:00:47.068997 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4fft\" (UniqueName: \"kubernetes.io/projected/9944885a-bc84-4efb-b0b5-6d850a4df958-kube-api-access-k4fft\") on node \"crc\" DevicePath \"\"" Dec 06 14:00:47 crc kubenswrapper[4706]: I1206 14:00:47.453000 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_collect-profiles-29417160-mfxrw_9944885a-bc84-4efb-b0b5-6d850a4df958/collect-profiles/0.log" Dec 06 14:00:47 crc kubenswrapper[4706]: I1206 14:00:47.453122 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw" event={"ID":"9944885a-bc84-4efb-b0b5-6d850a4df958","Type":"ContainerDied","Data":"0d188d8a26e5ce71fe14436dd9c32d03358062685b4ee85332d526a68e028a66"} Dec 06 14:00:47 crc kubenswrapper[4706]: I1206 14:00:47.453177 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d188d8a26e5ce71fe14436dd9c32d03358062685b4ee85332d526a68e028a66" Dec 06 14:00:47 crc kubenswrapper[4706]: I1206 14:00:47.453218 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-mfxrw" Dec 06 14:00:52 crc kubenswrapper[4706]: I1206 14:00:52.047275 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 14:00:56 crc kubenswrapper[4706]: I1206 14:00:56.516672 4706 generic.go:334] "Generic (PLEG): container finished" podID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerID="f7e97b997ea0bad25bb3f00476367a46da5d7d8a15abcdcdefaa0039a4cc2282" exitCode=0 Dec 06 14:00:56 crc kubenswrapper[4706]: I1206 14:00:56.516750 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" event={"ID":"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b","Type":"ContainerDied","Data":"f7e97b997ea0bad25bb3f00476367a46da5d7d8a15abcdcdefaa0039a4cc2282"} Dec 06 14:00:56 crc kubenswrapper[4706]: I1206 14:00:56.517955 4706 scope.go:117] "RemoveContainer" containerID="f7e97b997ea0bad25bb3f00476367a46da5d7d8a15abcdcdefaa0039a4cc2282" Dec 06 14:00:57 crc kubenswrapper[4706]: I1206 14:00:57.527155 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-cvtpw_c2e4f1d6-67ff-4990-bcbf-13ad23dc511b/marketplace-operator/1.log" Dec 06 14:00:57 crc kubenswrapper[4706]: I1206 14:00:57.529247 4706 generic.go:334] "Generic (PLEG): container finished" podID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerID="2b2f8e405feea257f6aa6539c87eddc462ea5d5087897a2e2c648829cee9890b" exitCode=1 Dec 06 14:00:57 crc kubenswrapper[4706]: I1206 14:00:57.529423 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" event={"ID":"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b","Type":"ContainerDied","Data":"2b2f8e405feea257f6aa6539c87eddc462ea5d5087897a2e2c648829cee9890b"} Dec 06 14:00:57 crc kubenswrapper[4706]: I1206 14:00:57.529490 4706 scope.go:117] "RemoveContainer" containerID="f7e97b997ea0bad25bb3f00476367a46da5d7d8a15abcdcdefaa0039a4cc2282" Dec 06 14:00:57 crc kubenswrapper[4706]: I1206 14:00:57.530526 4706 scope.go:117] "RemoveContainer" containerID="2b2f8e405feea257f6aa6539c87eddc462ea5d5087897a2e2c648829cee9890b" Dec 06 14:00:57 crc kubenswrapper[4706]: E1206 14:00:57.530896 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-cvtpw_openshift-marketplace(c2e4f1d6-67ff-4990-bcbf-13ad23dc511b)\"" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" Dec 06 14:00:58 crc kubenswrapper[4706]: I1206 14:00:58.538906 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-cvtpw_c2e4f1d6-67ff-4990-bcbf-13ad23dc511b/marketplace-operator/1.log" Dec 06 14:00:59 crc kubenswrapper[4706]: I1206 14:00:59.566768 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 14:01:01 crc kubenswrapper[4706]: I1206 14:01:01.331639 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 14:01:01 crc kubenswrapper[4706]: I1206 14:01:01.786509 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 14:01:01 crc kubenswrapper[4706]: I1206 14:01:01.787366 4706 scope.go:117] "RemoveContainer" containerID="2b2f8e405feea257f6aa6539c87eddc462ea5d5087897a2e2c648829cee9890b" Dec 06 14:01:01 crc kubenswrapper[4706]: E1206 14:01:01.787692 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-cvtpw_openshift-marketplace(c2e4f1d6-67ff-4990-bcbf-13ad23dc511b)\"" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" Dec 06 14:01:01 crc kubenswrapper[4706]: I1206 14:01:01.788212 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 14:01:02 crc kubenswrapper[4706]: I1206 14:01:02.580196 4706 scope.go:117] "RemoveContainer" containerID="2b2f8e405feea257f6aa6539c87eddc462ea5d5087897a2e2c648829cee9890b" Dec 06 14:01:02 crc kubenswrapper[4706]: E1206 14:01:02.581240 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-cvtpw_openshift-marketplace(c2e4f1d6-67ff-4990-bcbf-13ad23dc511b)\"" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" Dec 06 14:01:03 crc kubenswrapper[4706]: I1206 14:01:03.523740 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.075864 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4"] Dec 06 14:01:04 crc kubenswrapper[4706]: E1206 14:01:04.076225 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9944885a-bc84-4efb-b0b5-6d850a4df958" containerName="collect-profiles" Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.076246 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="9944885a-bc84-4efb-b0b5-6d850a4df958" containerName="collect-profiles" Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.076483 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="9944885a-bc84-4efb-b0b5-6d850a4df958" containerName="collect-profiles" Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.077224 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4" Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.081810 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.082148 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.090693 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4"] Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.256382 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/243e058a-ed7a-4bea-afbb-fe019f84645e-config-volume\") pod \"collect-profiles-29417160-6nls4\" (UID: \"243e058a-ed7a-4bea-afbb-fe019f84645e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4" Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.256584 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/243e058a-ed7a-4bea-afbb-fe019f84645e-secret-volume\") pod \"collect-profiles-29417160-6nls4\" (UID: \"243e058a-ed7a-4bea-afbb-fe019f84645e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4" Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.256695 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bqk5\" (UniqueName: \"kubernetes.io/projected/243e058a-ed7a-4bea-afbb-fe019f84645e-kube-api-access-5bqk5\") pod \"collect-profiles-29417160-6nls4\" (UID: \"243e058a-ed7a-4bea-afbb-fe019f84645e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4" Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.357909 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/243e058a-ed7a-4bea-afbb-fe019f84645e-secret-volume\") pod \"collect-profiles-29417160-6nls4\" (UID: \"243e058a-ed7a-4bea-afbb-fe019f84645e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4" Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.357977 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bqk5\" (UniqueName: \"kubernetes.io/projected/243e058a-ed7a-4bea-afbb-fe019f84645e-kube-api-access-5bqk5\") pod \"collect-profiles-29417160-6nls4\" (UID: \"243e058a-ed7a-4bea-afbb-fe019f84645e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4" Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.358042 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/243e058a-ed7a-4bea-afbb-fe019f84645e-config-volume\") pod \"collect-profiles-29417160-6nls4\" (UID: \"243e058a-ed7a-4bea-afbb-fe019f84645e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4" Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.358877 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/243e058a-ed7a-4bea-afbb-fe019f84645e-config-volume\") pod \"collect-profiles-29417160-6nls4\" (UID: \"243e058a-ed7a-4bea-afbb-fe019f84645e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4" Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.369926 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/243e058a-ed7a-4bea-afbb-fe019f84645e-secret-volume\") pod \"collect-profiles-29417160-6nls4\" (UID: \"243e058a-ed7a-4bea-afbb-fe019f84645e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4" Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.378251 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bqk5\" (UniqueName: \"kubernetes.io/projected/243e058a-ed7a-4bea-afbb-fe019f84645e-kube-api-access-5bqk5\") pod \"collect-profiles-29417160-6nls4\" (UID: \"243e058a-ed7a-4bea-afbb-fe019f84645e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4" Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.410114 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4" Dec 06 14:01:04 crc kubenswrapper[4706]: I1206 14:01:04.666582 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4"] Dec 06 14:01:05 crc kubenswrapper[4706]: I1206 14:01:05.603187 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_collect-profiles-29417160-6nls4_243e058a-ed7a-4bea-afbb-fe019f84645e/collect-profiles/0.log" Dec 06 14:01:05 crc kubenswrapper[4706]: I1206 14:01:05.603535 4706 generic.go:334] "Generic (PLEG): container finished" podID="243e058a-ed7a-4bea-afbb-fe019f84645e" containerID="a470b973fe427633ac9180a19a5a8a3d9c72d1b9b7abffbabd6f45fda967df45" exitCode=1 Dec 06 14:01:05 crc kubenswrapper[4706]: I1206 14:01:05.603569 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4" event={"ID":"243e058a-ed7a-4bea-afbb-fe019f84645e","Type":"ContainerDied","Data":"a470b973fe427633ac9180a19a5a8a3d9c72d1b9b7abffbabd6f45fda967df45"} Dec 06 14:01:05 crc kubenswrapper[4706]: I1206 14:01:05.603600 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4" event={"ID":"243e058a-ed7a-4bea-afbb-fe019f84645e","Type":"ContainerStarted","Data":"704f4e0324bbb36909b721878d32c47413092b68ca4ddc8e8eb78fdfe1a749fa"} Dec 06 14:01:06 crc kubenswrapper[4706]: I1206 14:01:06.881114 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_collect-profiles-29417160-6nls4_243e058a-ed7a-4bea-afbb-fe019f84645e/collect-profiles/0.log" Dec 06 14:01:06 crc kubenswrapper[4706]: I1206 14:01:06.881659 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4" Dec 06 14:01:07 crc kubenswrapper[4706]: I1206 14:01:07.005720 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bqk5\" (UniqueName: \"kubernetes.io/projected/243e058a-ed7a-4bea-afbb-fe019f84645e-kube-api-access-5bqk5\") pod \"243e058a-ed7a-4bea-afbb-fe019f84645e\" (UID: \"243e058a-ed7a-4bea-afbb-fe019f84645e\") " Dec 06 14:01:07 crc kubenswrapper[4706]: I1206 14:01:07.005868 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/243e058a-ed7a-4bea-afbb-fe019f84645e-config-volume\") pod \"243e058a-ed7a-4bea-afbb-fe019f84645e\" (UID: \"243e058a-ed7a-4bea-afbb-fe019f84645e\") " Dec 06 14:01:07 crc kubenswrapper[4706]: I1206 14:01:07.006701 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/243e058a-ed7a-4bea-afbb-fe019f84645e-secret-volume\") pod \"243e058a-ed7a-4bea-afbb-fe019f84645e\" (UID: \"243e058a-ed7a-4bea-afbb-fe019f84645e\") " Dec 06 14:01:07 crc kubenswrapper[4706]: I1206 14:01:07.006976 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/243e058a-ed7a-4bea-afbb-fe019f84645e-config-volume" (OuterVolumeSpecName: "config-volume") pod "243e058a-ed7a-4bea-afbb-fe019f84645e" (UID: "243e058a-ed7a-4bea-afbb-fe019f84645e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:01:07 crc kubenswrapper[4706]: I1206 14:01:07.007436 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/243e058a-ed7a-4bea-afbb-fe019f84645e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:07 crc kubenswrapper[4706]: I1206 14:01:07.013881 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/243e058a-ed7a-4bea-afbb-fe019f84645e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "243e058a-ed7a-4bea-afbb-fe019f84645e" (UID: "243e058a-ed7a-4bea-afbb-fe019f84645e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:01:07 crc kubenswrapper[4706]: I1206 14:01:07.014272 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/243e058a-ed7a-4bea-afbb-fe019f84645e-kube-api-access-5bqk5" (OuterVolumeSpecName: "kube-api-access-5bqk5") pod "243e058a-ed7a-4bea-afbb-fe019f84645e" (UID: "243e058a-ed7a-4bea-afbb-fe019f84645e"). InnerVolumeSpecName "kube-api-access-5bqk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:01:07 crc kubenswrapper[4706]: I1206 14:01:07.108555 4706 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/243e058a-ed7a-4bea-afbb-fe019f84645e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:07 crc kubenswrapper[4706]: I1206 14:01:07.108616 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bqk5\" (UniqueName: \"kubernetes.io/projected/243e058a-ed7a-4bea-afbb-fe019f84645e-kube-api-access-5bqk5\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:07 crc kubenswrapper[4706]: I1206 14:01:07.621742 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_collect-profiles-29417160-6nls4_243e058a-ed7a-4bea-afbb-fe019f84645e/collect-profiles/0.log" Dec 06 14:01:07 crc kubenswrapper[4706]: I1206 14:01:07.622139 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4" event={"ID":"243e058a-ed7a-4bea-afbb-fe019f84645e","Type":"ContainerDied","Data":"704f4e0324bbb36909b721878d32c47413092b68ca4ddc8e8eb78fdfe1a749fa"} Dec 06 14:01:07 crc kubenswrapper[4706]: I1206 14:01:07.622170 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="704f4e0324bbb36909b721878d32c47413092b68ca4ddc8e8eb78fdfe1a749fa" Dec 06 14:01:07 crc kubenswrapper[4706]: I1206 14:01:07.622253 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-6nls4" Dec 06 14:01:13 crc kubenswrapper[4706]: I1206 14:01:13.512304 4706 scope.go:117] "RemoveContainer" containerID="2b2f8e405feea257f6aa6539c87eddc462ea5d5087897a2e2c648829cee9890b" Dec 06 14:01:14 crc kubenswrapper[4706]: I1206 14:01:14.669332 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-cvtpw_c2e4f1d6-67ff-4990-bcbf-13ad23dc511b/marketplace-operator/2.log" Dec 06 14:01:14 crc kubenswrapper[4706]: I1206 14:01:14.670659 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-cvtpw_c2e4f1d6-67ff-4990-bcbf-13ad23dc511b/marketplace-operator/1.log" Dec 06 14:01:14 crc kubenswrapper[4706]: I1206 14:01:14.670725 4706 generic.go:334] "Generic (PLEG): container finished" podID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerID="f5c9f2be1c7a7f1de2b11119f16eefa6cf463a1257f5e8e9647c77d243bb5d04" exitCode=1 Dec 06 14:01:14 crc kubenswrapper[4706]: I1206 14:01:14.670770 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" event={"ID":"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b","Type":"ContainerDied","Data":"f5c9f2be1c7a7f1de2b11119f16eefa6cf463a1257f5e8e9647c77d243bb5d04"} Dec 06 14:01:14 crc kubenswrapper[4706]: I1206 14:01:14.670811 4706 scope.go:117] "RemoveContainer" containerID="2b2f8e405feea257f6aa6539c87eddc462ea5d5087897a2e2c648829cee9890b" Dec 06 14:01:14 crc kubenswrapper[4706]: I1206 14:01:14.671432 4706 scope.go:117] "RemoveContainer" containerID="f5c9f2be1c7a7f1de2b11119f16eefa6cf463a1257f5e8e9647c77d243bb5d04" Dec 06 14:01:14 crc kubenswrapper[4706]: E1206 14:01:14.671747 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-cvtpw_openshift-marketplace(c2e4f1d6-67ff-4990-bcbf-13ad23dc511b)\"" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" Dec 06 14:01:15 crc kubenswrapper[4706]: I1206 14:01:15.678945 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-cvtpw_c2e4f1d6-67ff-4990-bcbf-13ad23dc511b/marketplace-operator/2.log" Dec 06 14:01:21 crc kubenswrapper[4706]: I1206 14:01:21.786531 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 14:01:21 crc kubenswrapper[4706]: I1206 14:01:21.788435 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 14:01:21 crc kubenswrapper[4706]: I1206 14:01:21.790598 4706 scope.go:117] "RemoveContainer" containerID="f5c9f2be1c7a7f1de2b11119f16eefa6cf463a1257f5e8e9647c77d243bb5d04" Dec 06 14:01:21 crc kubenswrapper[4706]: E1206 14:01:21.790964 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-cvtpw_openshift-marketplace(c2e4f1d6-67ff-4990-bcbf-13ad23dc511b)\"" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" Dec 06 14:01:23 crc kubenswrapper[4706]: I1206 14:01:23.505763 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 14:01:34 crc kubenswrapper[4706]: I1206 14:01:34.511001 4706 scope.go:117] "RemoveContainer" containerID="f5c9f2be1c7a7f1de2b11119f16eefa6cf463a1257f5e8e9647c77d243bb5d04" Dec 06 14:01:34 crc kubenswrapper[4706]: I1206 14:01:34.799358 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-cvtpw_c2e4f1d6-67ff-4990-bcbf-13ad23dc511b/marketplace-operator/2.log" Dec 06 14:01:34 crc kubenswrapper[4706]: I1206 14:01:34.799436 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" event={"ID":"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b","Type":"ContainerStarted","Data":"8b02855834aaad547d00e96eea1d4ee7acf8d6c2c732f0d642b9138b1de49b2b"} Dec 06 14:01:34 crc kubenswrapper[4706]: I1206 14:01:34.799891 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 14:01:34 crc kubenswrapper[4706]: I1206 14:01:34.800669 4706 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cvtpw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 06 14:01:34 crc kubenswrapper[4706]: I1206 14:01:34.800720 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 06 14:01:35 crc kubenswrapper[4706]: I1206 14:01:35.807156 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.528795 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kslt9"] Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.532098 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kslt9" podUID="1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" containerName="registry-server" containerID="cri-o://87ffbb89d97e84a1396792279ee8676f1771b30d9cdafddae97a066f3965e14d" gracePeriod=30 Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.535859 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hjkcq"] Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.536129 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hjkcq" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" containerName="registry-server" containerID="cri-o://0e79497c67181657b28b5995c7609d75815eb1bddd2ee6261744d3cfdcd41f01" gracePeriod=30 Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.545579 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zhlvc"] Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.545872 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zhlvc" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" containerName="registry-server" containerID="cri-o://e3ff9ae87bec78ac5f3199a1b0baa2fbbcf7b3f5f0a3bd99d9ae29b94e2c2650" gracePeriod=30 Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.552193 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cvtpw"] Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.553550 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerName="marketplace-operator" containerID="cri-o://8b02855834aaad547d00e96eea1d4ee7acf8d6c2c732f0d642b9138b1de49b2b" gracePeriod=30 Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.571609 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdw9k"] Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.571899 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wdw9k" podUID="906c227f-a713-49d2-acfd-633646b6e0da" containerName="registry-server" containerID="cri-o://2889d570d3639df1831a6543ad6ef2320dbe0c6d06bbce50397a22fc9d2582de" gracePeriod=30 Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.576586 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sgfd9"] Dec 06 14:01:38 crc kubenswrapper[4706]: E1206 14:01:38.576828 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243e058a-ed7a-4bea-afbb-fe019f84645e" containerName="collect-profiles" Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.576839 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="243e058a-ed7a-4bea-afbb-fe019f84645e" containerName="collect-profiles" Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.576939 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="243e058a-ed7a-4bea-afbb-fe019f84645e" containerName="collect-profiles" Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.577317 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sgfd9" Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.581695 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x5pm6"] Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.581903 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x5pm6" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" containerName="registry-server" containerID="cri-o://6f740d3590d0a3cf15468d637d45fb68f3ecd5bdcf66d802d6cadc1bf62a80bc" gracePeriod=30 Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.598028 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sgfd9"] Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.658068 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkjrg\" (UniqueName: \"kubernetes.io/projected/92920365-098a-442a-87de-91aa1479ff1d-kube-api-access-fkjrg\") pod \"marketplace-operator-79b997595-sgfd9\" (UID: \"92920365-098a-442a-87de-91aa1479ff1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgfd9" Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.658507 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/92920365-098a-442a-87de-91aa1479ff1d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sgfd9\" (UID: \"92920365-098a-442a-87de-91aa1479ff1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgfd9" Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.658882 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/92920365-098a-442a-87de-91aa1479ff1d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sgfd9\" (UID: \"92920365-098a-442a-87de-91aa1479ff1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgfd9" Dec 06 14:01:38 crc kubenswrapper[4706]: E1206 14:01:38.734829 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0e79497c67181657b28b5995c7609d75815eb1bddd2ee6261744d3cfdcd41f01 is running failed: container process not found" containerID="0e79497c67181657b28b5995c7609d75815eb1bddd2ee6261744d3cfdcd41f01" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 14:01:38 crc kubenswrapper[4706]: E1206 14:01:38.735213 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0e79497c67181657b28b5995c7609d75815eb1bddd2ee6261744d3cfdcd41f01 is running failed: container process not found" containerID="0e79497c67181657b28b5995c7609d75815eb1bddd2ee6261744d3cfdcd41f01" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 14:01:38 crc kubenswrapper[4706]: E1206 14:01:38.735499 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0e79497c67181657b28b5995c7609d75815eb1bddd2ee6261744d3cfdcd41f01 is running failed: container process not found" containerID="0e79497c67181657b28b5995c7609d75815eb1bddd2ee6261744d3cfdcd41f01" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 14:01:38 crc kubenswrapper[4706]: E1206 14:01:38.735542 4706 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0e79497c67181657b28b5995c7609d75815eb1bddd2ee6261744d3cfdcd41f01 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-hjkcq" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" containerName="registry-server" Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.760784 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkjrg\" (UniqueName: \"kubernetes.io/projected/92920365-098a-442a-87de-91aa1479ff1d-kube-api-access-fkjrg\") pod \"marketplace-operator-79b997595-sgfd9\" (UID: \"92920365-098a-442a-87de-91aa1479ff1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgfd9" Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.760862 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/92920365-098a-442a-87de-91aa1479ff1d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sgfd9\" (UID: \"92920365-098a-442a-87de-91aa1479ff1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgfd9" Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.760897 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/92920365-098a-442a-87de-91aa1479ff1d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sgfd9\" (UID: \"92920365-098a-442a-87de-91aa1479ff1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgfd9" Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.762818 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/92920365-098a-442a-87de-91aa1479ff1d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sgfd9\" (UID: \"92920365-098a-442a-87de-91aa1479ff1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgfd9" Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.766714 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/92920365-098a-442a-87de-91aa1479ff1d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sgfd9\" (UID: \"92920365-098a-442a-87de-91aa1479ff1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgfd9" Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.782978 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkjrg\" (UniqueName: \"kubernetes.io/projected/92920365-098a-442a-87de-91aa1479ff1d-kube-api-access-fkjrg\") pod \"marketplace-operator-79b997595-sgfd9\" (UID: \"92920365-098a-442a-87de-91aa1479ff1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgfd9" Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.822899 4706 generic.go:334] "Generic (PLEG): container finished" podID="1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" containerID="87ffbb89d97e84a1396792279ee8676f1771b30d9cdafddae97a066f3965e14d" exitCode=0 Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.822958 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kslt9" event={"ID":"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2","Type":"ContainerDied","Data":"87ffbb89d97e84a1396792279ee8676f1771b30d9cdafddae97a066f3965e14d"} Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.825517 4706 generic.go:334] "Generic (PLEG): container finished" podID="906c227f-a713-49d2-acfd-633646b6e0da" containerID="2889d570d3639df1831a6543ad6ef2320dbe0c6d06bbce50397a22fc9d2582de" exitCode=0 Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.825558 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdw9k" event={"ID":"906c227f-a713-49d2-acfd-633646b6e0da","Type":"ContainerDied","Data":"2889d570d3639df1831a6543ad6ef2320dbe0c6d06bbce50397a22fc9d2582de"} Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.837220 4706 generic.go:334] "Generic (PLEG): container finished" podID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" containerID="0e79497c67181657b28b5995c7609d75815eb1bddd2ee6261744d3cfdcd41f01" exitCode=0 Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.837314 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjkcq" event={"ID":"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5","Type":"ContainerDied","Data":"0e79497c67181657b28b5995c7609d75815eb1bddd2ee6261744d3cfdcd41f01"} Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.839423 4706 generic.go:334] "Generic (PLEG): container finished" podID="d12f4107-0d89-4d3c-958f-7355fbd27575" containerID="e3ff9ae87bec78ac5f3199a1b0baa2fbbcf7b3f5f0a3bd99d9ae29b94e2c2650" exitCode=0 Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.839485 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zhlvc" event={"ID":"d12f4107-0d89-4d3c-958f-7355fbd27575","Type":"ContainerDied","Data":"e3ff9ae87bec78ac5f3199a1b0baa2fbbcf7b3f5f0a3bd99d9ae29b94e2c2650"} Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.841332 4706 generic.go:334] "Generic (PLEG): container finished" podID="34cf831e-196b-46c2-a2a1-d985cdfd8861" containerID="6f740d3590d0a3cf15468d637d45fb68f3ecd5bdcf66d802d6cadc1bf62a80bc" exitCode=0 Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.841381 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5pm6" event={"ID":"34cf831e-196b-46c2-a2a1-d985cdfd8861","Type":"ContainerDied","Data":"6f740d3590d0a3cf15468d637d45fb68f3ecd5bdcf66d802d6cadc1bf62a80bc"} Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.843957 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-cvtpw_c2e4f1d6-67ff-4990-bcbf-13ad23dc511b/marketplace-operator/2.log" Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.844004 4706 generic.go:334] "Generic (PLEG): container finished" podID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerID="8b02855834aaad547d00e96eea1d4ee7acf8d6c2c732f0d642b9138b1de49b2b" exitCode=0 Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.844027 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" event={"ID":"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b","Type":"ContainerDied","Data":"8b02855834aaad547d00e96eea1d4ee7acf8d6c2c732f0d642b9138b1de49b2b"} Dec 06 14:01:38 crc kubenswrapper[4706]: I1206 14:01:38.844053 4706 scope.go:117] "RemoveContainer" containerID="f5c9f2be1c7a7f1de2b11119f16eefa6cf463a1257f5e8e9647c77d243bb5d04" Dec 06 14:01:38 crc kubenswrapper[4706]: E1206 14:01:38.887107 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 87ffbb89d97e84a1396792279ee8676f1771b30d9cdafddae97a066f3965e14d is running failed: container process not found" containerID="87ffbb89d97e84a1396792279ee8676f1771b30d9cdafddae97a066f3965e14d" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 14:01:38 crc kubenswrapper[4706]: E1206 14:01:38.887528 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 87ffbb89d97e84a1396792279ee8676f1771b30d9cdafddae97a066f3965e14d is running failed: container process not found" containerID="87ffbb89d97e84a1396792279ee8676f1771b30d9cdafddae97a066f3965e14d" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 14:01:38 crc kubenswrapper[4706]: E1206 14:01:38.887836 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 87ffbb89d97e84a1396792279ee8676f1771b30d9cdafddae97a066f3965e14d is running failed: container process not found" containerID="87ffbb89d97e84a1396792279ee8676f1771b30d9cdafddae97a066f3965e14d" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 14:01:38 crc kubenswrapper[4706]: E1206 14:01:38.887873 4706 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 87ffbb89d97e84a1396792279ee8676f1771b30d9cdafddae97a066f3965e14d is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-kslt9" podUID="1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" containerName="registry-server" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.005769 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sgfd9" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.021620 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hjkcq" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.029927 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zhlvc" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.037677 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5pm6" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.045802 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kslt9" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.056693 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdw9k" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.071749 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.171931 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34cf831e-196b-46c2-a2a1-d985cdfd8861-utilities\") pod \"34cf831e-196b-46c2-a2a1-d985cdfd8861\" (UID: \"34cf831e-196b-46c2-a2a1-d985cdfd8861\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.172001 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p4zf\" (UniqueName: \"kubernetes.io/projected/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-kube-api-access-6p4zf\") pod \"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2\" (UID: \"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.172033 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-catalog-content\") pod \"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2\" (UID: \"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.172058 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzjxf\" (UniqueName: \"kubernetes.io/projected/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-kube-api-access-tzjxf\") pod \"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b\" (UID: \"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.172075 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbp2v\" (UniqueName: \"kubernetes.io/projected/906c227f-a713-49d2-acfd-633646b6e0da-kube-api-access-qbp2v\") pod \"906c227f-a713-49d2-acfd-633646b6e0da\" (UID: \"906c227f-a713-49d2-acfd-633646b6e0da\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.172096 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34cf831e-196b-46c2-a2a1-d985cdfd8861-catalog-content\") pod \"34cf831e-196b-46c2-a2a1-d985cdfd8861\" (UID: \"34cf831e-196b-46c2-a2a1-d985cdfd8861\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.172120 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-utilities\") pod \"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2\" (UID: \"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.172139 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d12f4107-0d89-4d3c-958f-7355fbd27575-utilities\") pod \"d12f4107-0d89-4d3c-958f-7355fbd27575\" (UID: \"d12f4107-0d89-4d3c-958f-7355fbd27575\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.172157 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-trusted-ca\") pod \"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b\" (UID: \"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.172173 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvbhh\" (UniqueName: \"kubernetes.io/projected/34cf831e-196b-46c2-a2a1-d985cdfd8861-kube-api-access-fvbhh\") pod \"34cf831e-196b-46c2-a2a1-d985cdfd8861\" (UID: \"34cf831e-196b-46c2-a2a1-d985cdfd8861\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.172204 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906c227f-a713-49d2-acfd-633646b6e0da-catalog-content\") pod \"906c227f-a713-49d2-acfd-633646b6e0da\" (UID: \"906c227f-a713-49d2-acfd-633646b6e0da\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.172224 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-operator-metrics\") pod \"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b\" (UID: \"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.172250 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-catalog-content\") pod \"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5\" (UID: \"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.172268 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-utilities\") pod \"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5\" (UID: \"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.172286 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxrzw\" (UniqueName: \"kubernetes.io/projected/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-kube-api-access-fxrzw\") pod \"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5\" (UID: \"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.172326 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d12f4107-0d89-4d3c-958f-7355fbd27575-catalog-content\") pod \"d12f4107-0d89-4d3c-958f-7355fbd27575\" (UID: \"d12f4107-0d89-4d3c-958f-7355fbd27575\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.172347 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906c227f-a713-49d2-acfd-633646b6e0da-utilities\") pod \"906c227f-a713-49d2-acfd-633646b6e0da\" (UID: \"906c227f-a713-49d2-acfd-633646b6e0da\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.172379 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w4bj\" (UniqueName: \"kubernetes.io/projected/d12f4107-0d89-4d3c-958f-7355fbd27575-kube-api-access-9w4bj\") pod \"d12f4107-0d89-4d3c-958f-7355fbd27575\" (UID: \"d12f4107-0d89-4d3c-958f-7355fbd27575\") " Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.173934 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34cf831e-196b-46c2-a2a1-d985cdfd8861-utilities" (OuterVolumeSpecName: "utilities") pod "34cf831e-196b-46c2-a2a1-d985cdfd8861" (UID: "34cf831e-196b-46c2-a2a1-d985cdfd8861"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.178589 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/906c227f-a713-49d2-acfd-633646b6e0da-utilities" (OuterVolumeSpecName: "utilities") pod "906c227f-a713-49d2-acfd-633646b6e0da" (UID: "906c227f-a713-49d2-acfd-633646b6e0da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.178890 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-utilities" (OuterVolumeSpecName: "utilities") pod "93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" (UID: "93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.180408 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" (UID: "c2e4f1d6-67ff-4990-bcbf-13ad23dc511b"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.180533 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d12f4107-0d89-4d3c-958f-7355fbd27575-utilities" (OuterVolumeSpecName: "utilities") pod "d12f4107-0d89-4d3c-958f-7355fbd27575" (UID: "d12f4107-0d89-4d3c-958f-7355fbd27575"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.180885 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-utilities" (OuterVolumeSpecName: "utilities") pod "1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" (UID: "1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.182432 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d12f4107-0d89-4d3c-958f-7355fbd27575-kube-api-access-9w4bj" (OuterVolumeSpecName: "kube-api-access-9w4bj") pod "d12f4107-0d89-4d3c-958f-7355fbd27575" (UID: "d12f4107-0d89-4d3c-958f-7355fbd27575"). InnerVolumeSpecName "kube-api-access-9w4bj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.187621 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-kube-api-access-fxrzw" (OuterVolumeSpecName: "kube-api-access-fxrzw") pod "93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" (UID: "93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5"). InnerVolumeSpecName "kube-api-access-fxrzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.190012 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-kube-api-access-6p4zf" (OuterVolumeSpecName: "kube-api-access-6p4zf") pod "1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" (UID: "1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2"). InnerVolumeSpecName "kube-api-access-6p4zf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.199621 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/906c227f-a713-49d2-acfd-633646b6e0da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "906c227f-a713-49d2-acfd-633646b6e0da" (UID: "906c227f-a713-49d2-acfd-633646b6e0da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.205392 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" (UID: "c2e4f1d6-67ff-4990-bcbf-13ad23dc511b"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.228141 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/906c227f-a713-49d2-acfd-633646b6e0da-kube-api-access-qbp2v" (OuterVolumeSpecName: "kube-api-access-qbp2v") pod "906c227f-a713-49d2-acfd-633646b6e0da" (UID: "906c227f-a713-49d2-acfd-633646b6e0da"). InnerVolumeSpecName "kube-api-access-qbp2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.228205 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34cf831e-196b-46c2-a2a1-d985cdfd8861-kube-api-access-fvbhh" (OuterVolumeSpecName: "kube-api-access-fvbhh") pod "34cf831e-196b-46c2-a2a1-d985cdfd8861" (UID: "34cf831e-196b-46c2-a2a1-d985cdfd8861"). InnerVolumeSpecName "kube-api-access-fvbhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.228815 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sgfd9"] Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.230493 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-kube-api-access-tzjxf" (OuterVolumeSpecName: "kube-api-access-tzjxf") pod "c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" (UID: "c2e4f1d6-67ff-4990-bcbf-13ad23dc511b"). InnerVolumeSpecName "kube-api-access-tzjxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.240317 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" (UID: "93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.259426 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d12f4107-0d89-4d3c-958f-7355fbd27575-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d12f4107-0d89-4d3c-958f-7355fbd27575" (UID: "d12f4107-0d89-4d3c-958f-7355fbd27575"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.274175 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906c227f-a713-49d2-acfd-633646b6e0da-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.274204 4706 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.274217 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.274226 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.274235 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxrzw\" (UniqueName: \"kubernetes.io/projected/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5-kube-api-access-fxrzw\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.274244 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d12f4107-0d89-4d3c-958f-7355fbd27575-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.274251 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906c227f-a713-49d2-acfd-633646b6e0da-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.274259 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w4bj\" (UniqueName: \"kubernetes.io/projected/d12f4107-0d89-4d3c-958f-7355fbd27575-kube-api-access-9w4bj\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.274267 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34cf831e-196b-46c2-a2a1-d985cdfd8861-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.274275 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p4zf\" (UniqueName: \"kubernetes.io/projected/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-kube-api-access-6p4zf\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.274285 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzjxf\" (UniqueName: \"kubernetes.io/projected/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-kube-api-access-tzjxf\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.274294 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbp2v\" (UniqueName: \"kubernetes.io/projected/906c227f-a713-49d2-acfd-633646b6e0da-kube-api-access-qbp2v\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.274301 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.274322 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d12f4107-0d89-4d3c-958f-7355fbd27575-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.274330 4706 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.274338 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvbhh\" (UniqueName: \"kubernetes.io/projected/34cf831e-196b-46c2-a2a1-d985cdfd8861-kube-api-access-fvbhh\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.276601 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" (UID: "1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.347180 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34cf831e-196b-46c2-a2a1-d985cdfd8861-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34cf831e-196b-46c2-a2a1-d985cdfd8861" (UID: "34cf831e-196b-46c2-a2a1-d985cdfd8861"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.375034 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.375064 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34cf831e-196b-46c2-a2a1-d985cdfd8861-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.851670 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kslt9" event={"ID":"1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2","Type":"ContainerDied","Data":"dcfdb05768a03816b2f74b6417276cb768e3940493e24a62398056ff02337b6d"} Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.852003 4706 scope.go:117] "RemoveContainer" containerID="87ffbb89d97e84a1396792279ee8676f1771b30d9cdafddae97a066f3965e14d" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.851720 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kslt9" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.855029 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdw9k" event={"ID":"906c227f-a713-49d2-acfd-633646b6e0da","Type":"ContainerDied","Data":"7124ce5e15f7aa7359e7e501561a5f3f5982a75a865ea5ebe4dc3a29576f6a30"} Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.855131 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdw9k" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.859543 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hjkcq" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.859588 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjkcq" event={"ID":"93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5","Type":"ContainerDied","Data":"badef2b00d9a2dc3a88364f683c49374cd32ae3dd24770c54b8f16b457a3c0b0"} Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.863507 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zhlvc" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.863580 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zhlvc" event={"ID":"d12f4107-0d89-4d3c-958f-7355fbd27575","Type":"ContainerDied","Data":"79e0531c01ab0055b715bce21638bcba539340cee7ae422b15aff937e9f69799"} Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.866986 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5pm6" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.866991 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5pm6" event={"ID":"34cf831e-196b-46c2-a2a1-d985cdfd8861","Type":"ContainerDied","Data":"8aa6bb249443c49e315b16d56261c28ead42a575c350e18aa8a1b60072b9bbd4"} Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.869589 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.869624 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cvtpw" event={"ID":"c2e4f1d6-67ff-4990-bcbf-13ad23dc511b","Type":"ContainerDied","Data":"1e8fda098971671008f1b8749d30e747e0444484738057bda313edb0af94698a"} Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.872082 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sgfd9" event={"ID":"92920365-098a-442a-87de-91aa1479ff1d","Type":"ContainerStarted","Data":"bfc4c499e8efd1a6ea43e3b2c48df7afe76207c75cd13c11556f767958f98488"} Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.872144 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sgfd9" event={"ID":"92920365-098a-442a-87de-91aa1479ff1d","Type":"ContainerStarted","Data":"b3a68195d79d32d3183bf134ff9b8d462ece4a531e968f8aa1a80d6fb242dc06"} Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.872711 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-sgfd9" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.877400 4706 scope.go:117] "RemoveContainer" containerID="8571919fedef679e1130f6f1ddbad192d56d022c63105ccf1a58e8843a77353e" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.877823 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-sgfd9" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.895560 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-sgfd9" podStartSLOduration=1.895537209 podStartE2EDuration="1.895537209s" podCreationTimestamp="2025-12-06 14:01:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:01:39.889285952 +0000 UTC m=+342.725078829" watchObservedRunningTime="2025-12-06 14:01:39.895537209 +0000 UTC m=+342.731330066" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.912665 4706 scope.go:117] "RemoveContainer" containerID="7f974c473ff680877f09f18406c1f699e8c3f551fd1bafb7cada66e04cb0bfa2" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.936341 4706 scope.go:117] "RemoveContainer" containerID="2889d570d3639df1831a6543ad6ef2320dbe0c6d06bbce50397a22fc9d2582de" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.949973 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zhlvc"] Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.952207 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zhlvc"] Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.960438 4706 scope.go:117] "RemoveContainer" containerID="19d38017e092899a7cfd0aa5a92e4bd96a09ce9a9759dc303fb55cbd529fceba" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.961780 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdw9k"] Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.969889 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdw9k"] Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.974858 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hjkcq"] Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.976472 4706 scope.go:117] "RemoveContainer" containerID="91195cfe1f00c83e86251aa226042c619b7f0b3ed291c91a38e02209edcc50ce" Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.979416 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hjkcq"] Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.986392 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kslt9"] Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.991465 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kslt9"] Dec 06 14:01:39 crc kubenswrapper[4706]: I1206 14:01:39.998397 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cvtpw"] Dec 06 14:01:40 crc kubenswrapper[4706]: I1206 14:01:40.000131 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cvtpw"] Dec 06 14:01:40 crc kubenswrapper[4706]: I1206 14:01:40.003156 4706 scope.go:117] "RemoveContainer" containerID="0e79497c67181657b28b5995c7609d75815eb1bddd2ee6261744d3cfdcd41f01" Dec 06 14:01:40 crc kubenswrapper[4706]: I1206 14:01:40.007660 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x5pm6"] Dec 06 14:01:40 crc kubenswrapper[4706]: I1206 14:01:40.016792 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x5pm6"] Dec 06 14:01:40 crc kubenswrapper[4706]: I1206 14:01:40.021969 4706 scope.go:117] "RemoveContainer" containerID="08397677c6ee59183c74d79ad34b5b042c1caba72fb7831df4e72a09c6472615" Dec 06 14:01:40 crc kubenswrapper[4706]: I1206 14:01:40.032617 4706 scope.go:117] "RemoveContainer" containerID="aac2689774446cc2f9ba1997a613d4e735786174daa0e2e0135c7f73aaeeb65d" Dec 06 14:01:40 crc kubenswrapper[4706]: I1206 14:01:40.043950 4706 scope.go:117] "RemoveContainer" containerID="e3ff9ae87bec78ac5f3199a1b0baa2fbbcf7b3f5f0a3bd99d9ae29b94e2c2650" Dec 06 14:01:40 crc kubenswrapper[4706]: I1206 14:01:40.055835 4706 scope.go:117] "RemoveContainer" containerID="8f77e7cfa55d0de9299ca6f6184f7d4a3cb8ee885a51d505f2a8fef78438c9dc" Dec 06 14:01:40 crc kubenswrapper[4706]: I1206 14:01:40.068262 4706 scope.go:117] "RemoveContainer" containerID="6447594ac6b38ab5e2a58a1551c4b2243c7ae35e0aae2468cb919993d61123c4" Dec 06 14:01:40 crc kubenswrapper[4706]: I1206 14:01:40.081905 4706 scope.go:117] "RemoveContainer" containerID="6f740d3590d0a3cf15468d637d45fb68f3ecd5bdcf66d802d6cadc1bf62a80bc" Dec 06 14:01:40 crc kubenswrapper[4706]: I1206 14:01:40.093126 4706 scope.go:117] "RemoveContainer" containerID="e192b62048f70495e09c722bfaf3f7bd4724fe91c292baeb5a3cf18618672522" Dec 06 14:01:40 crc kubenswrapper[4706]: I1206 14:01:40.108967 4706 scope.go:117] "RemoveContainer" containerID="bbe5490c2e1844ce1d4095ad4ee7bcc4dc2548492ed6bc2933f96418e8667b5a" Dec 06 14:01:40 crc kubenswrapper[4706]: I1206 14:01:40.122416 4706 scope.go:117] "RemoveContainer" containerID="8b02855834aaad547d00e96eea1d4ee7acf8d6c2c732f0d642b9138b1de49b2b" Dec 06 14:01:41 crc kubenswrapper[4706]: I1206 14:01:41.522012 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" path="/var/lib/kubelet/pods/1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2/volumes" Dec 06 14:01:41 crc kubenswrapper[4706]: I1206 14:01:41.522931 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" path="/var/lib/kubelet/pods/34cf831e-196b-46c2-a2a1-d985cdfd8861/volumes" Dec 06 14:01:41 crc kubenswrapper[4706]: I1206 14:01:41.524506 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="906c227f-a713-49d2-acfd-633646b6e0da" path="/var/lib/kubelet/pods/906c227f-a713-49d2-acfd-633646b6e0da/volumes" Dec 06 14:01:41 crc kubenswrapper[4706]: I1206 14:01:41.525556 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" path="/var/lib/kubelet/pods/93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5/volumes" Dec 06 14:01:41 crc kubenswrapper[4706]: I1206 14:01:41.526214 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" path="/var/lib/kubelet/pods/c2e4f1d6-67ff-4990-bcbf-13ad23dc511b/volumes" Dec 06 14:01:41 crc kubenswrapper[4706]: I1206 14:01:41.527062 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" path="/var/lib/kubelet/pods/d12f4107-0d89-4d3c-958f-7355fbd27575/volumes" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.019841 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx"] Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020323 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="906c227f-a713-49d2-acfd-633646b6e0da" containerName="registry-server" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020334 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="906c227f-a713-49d2-acfd-633646b6e0da" containerName="registry-server" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020346 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerName="marketplace-operator" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020353 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerName="marketplace-operator" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020360 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" containerName="registry-server" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020366 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" containerName="registry-server" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020375 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" containerName="registry-server" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020381 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" containerName="registry-server" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020390 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" containerName="extract-content" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020395 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" containerName="extract-content" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020402 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" containerName="extract-content" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020408 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" containerName="extract-content" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020416 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" containerName="registry-server" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020421 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" containerName="registry-server" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020430 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" containerName="extract-utilities" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020436 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" containerName="extract-utilities" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020444 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" containerName="extract-utilities" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020450 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" containerName="extract-utilities" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020457 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" containerName="extract-utilities" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020480 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" containerName="extract-utilities" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020488 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" containerName="extract-content" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020494 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" containerName="extract-content" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020503 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerName="marketplace-operator" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020509 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerName="marketplace-operator" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020515 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerName="marketplace-operator" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020520 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerName="marketplace-operator" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020529 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="906c227f-a713-49d2-acfd-633646b6e0da" containerName="extract-content" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020534 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="906c227f-a713-49d2-acfd-633646b6e0da" containerName="extract-content" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020542 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" containerName="extract-content" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020548 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" containerName="extract-content" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020557 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" containerName="registry-server" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020562 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" containerName="registry-server" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020568 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerName="marketplace-operator" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020574 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerName="marketplace-operator" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020582 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="906c227f-a713-49d2-acfd-633646b6e0da" containerName="extract-utilities" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020588 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="906c227f-a713-49d2-acfd-633646b6e0da" containerName="extract-utilities" Dec 06 14:01:44 crc kubenswrapper[4706]: E1206 14:01:44.020596 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" containerName="extract-utilities" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020602 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" containerName="extract-utilities" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020684 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerName="marketplace-operator" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020694 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerName="marketplace-operator" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020701 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a4b3ae-ccf6-4a2a-b2a3-e025cf0231d5" containerName="registry-server" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020710 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="906c227f-a713-49d2-acfd-633646b6e0da" containerName="registry-server" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020718 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerName="marketplace-operator" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020725 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a2f9cd3-2137-4ca3-b67f-2dbc543e49c2" containerName="registry-server" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020733 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d12f4107-0d89-4d3c-958f-7355fbd27575" containerName="registry-server" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.020742 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="34cf831e-196b-46c2-a2a1-d985cdfd8861" containerName="registry-server" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.021068 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.022648 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.023335 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.032246 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx"] Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.131517 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnkkg\" (UniqueName: \"kubernetes.io/projected/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-kube-api-access-fnkkg\") pod \"collect-profiles-29417160-7k5tx\" (UID: \"36de8e46-6ee6-4de4-bf0c-7ce0238dd655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.131660 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-config-volume\") pod \"collect-profiles-29417160-7k5tx\" (UID: \"36de8e46-6ee6-4de4-bf0c-7ce0238dd655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.131813 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-secret-volume\") pod \"collect-profiles-29417160-7k5tx\" (UID: \"36de8e46-6ee6-4de4-bf0c-7ce0238dd655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.233167 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnkkg\" (UniqueName: \"kubernetes.io/projected/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-kube-api-access-fnkkg\") pod \"collect-profiles-29417160-7k5tx\" (UID: \"36de8e46-6ee6-4de4-bf0c-7ce0238dd655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.233275 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-config-volume\") pod \"collect-profiles-29417160-7k5tx\" (UID: \"36de8e46-6ee6-4de4-bf0c-7ce0238dd655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.233378 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-secret-volume\") pod \"collect-profiles-29417160-7k5tx\" (UID: \"36de8e46-6ee6-4de4-bf0c-7ce0238dd655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.236115 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-config-volume\") pod \"collect-profiles-29417160-7k5tx\" (UID: \"36de8e46-6ee6-4de4-bf0c-7ce0238dd655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.242916 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-secret-volume\") pod \"collect-profiles-29417160-7k5tx\" (UID: \"36de8e46-6ee6-4de4-bf0c-7ce0238dd655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.255708 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnkkg\" (UniqueName: \"kubernetes.io/projected/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-kube-api-access-fnkkg\") pod \"collect-profiles-29417160-7k5tx\" (UID: \"36de8e46-6ee6-4de4-bf0c-7ce0238dd655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.346817 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.716374 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx"] Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.906550 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" event={"ID":"36de8e46-6ee6-4de4-bf0c-7ce0238dd655","Type":"ContainerStarted","Data":"b07d57470d0efedcc0d75219974b69ac0c1c32348b37d6b8d701cb35a7c87e68"} Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.906863 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" event={"ID":"36de8e46-6ee6-4de4-bf0c-7ce0238dd655","Type":"ContainerStarted","Data":"5f31b47bfc220648d39d7bfcdcdeb011b456d71d02d421c2d2e7bf944a64eb95"} Dec 06 14:01:44 crc kubenswrapper[4706]: I1206 14:01:44.919389 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" podStartSLOduration=0.919376284 podStartE2EDuration="919.376284ms" podCreationTimestamp="2025-12-06 14:01:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:01:44.917566406 +0000 UTC m=+347.753359263" watchObservedRunningTime="2025-12-06 14:01:44.919376284 +0000 UTC m=+347.755169141" Dec 06 14:01:45 crc kubenswrapper[4706]: I1206 14:01:45.913797 4706 generic.go:334] "Generic (PLEG): container finished" podID="36de8e46-6ee6-4de4-bf0c-7ce0238dd655" containerID="b07d57470d0efedcc0d75219974b69ac0c1c32348b37d6b8d701cb35a7c87e68" exitCode=0 Dec 06 14:01:45 crc kubenswrapper[4706]: I1206 14:01:45.913919 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" event={"ID":"36de8e46-6ee6-4de4-bf0c-7ce0238dd655","Type":"ContainerDied","Data":"b07d57470d0efedcc0d75219974b69ac0c1c32348b37d6b8d701cb35a7c87e68"} Dec 06 14:01:47 crc kubenswrapper[4706]: I1206 14:01:47.359422 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" Dec 06 14:01:47 crc kubenswrapper[4706]: I1206 14:01:47.477380 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-secret-volume\") pod \"36de8e46-6ee6-4de4-bf0c-7ce0238dd655\" (UID: \"36de8e46-6ee6-4de4-bf0c-7ce0238dd655\") " Dec 06 14:01:47 crc kubenswrapper[4706]: I1206 14:01:47.477473 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnkkg\" (UniqueName: \"kubernetes.io/projected/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-kube-api-access-fnkkg\") pod \"36de8e46-6ee6-4de4-bf0c-7ce0238dd655\" (UID: \"36de8e46-6ee6-4de4-bf0c-7ce0238dd655\") " Dec 06 14:01:47 crc kubenswrapper[4706]: I1206 14:01:47.477530 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-config-volume\") pod \"36de8e46-6ee6-4de4-bf0c-7ce0238dd655\" (UID: \"36de8e46-6ee6-4de4-bf0c-7ce0238dd655\") " Dec 06 14:01:47 crc kubenswrapper[4706]: I1206 14:01:47.478708 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-config-volume" (OuterVolumeSpecName: "config-volume") pod "36de8e46-6ee6-4de4-bf0c-7ce0238dd655" (UID: "36de8e46-6ee6-4de4-bf0c-7ce0238dd655"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:01:47 crc kubenswrapper[4706]: I1206 14:01:47.483932 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-kube-api-access-fnkkg" (OuterVolumeSpecName: "kube-api-access-fnkkg") pod "36de8e46-6ee6-4de4-bf0c-7ce0238dd655" (UID: "36de8e46-6ee6-4de4-bf0c-7ce0238dd655"). InnerVolumeSpecName "kube-api-access-fnkkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:01:47 crc kubenswrapper[4706]: I1206 14:01:47.484456 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "36de8e46-6ee6-4de4-bf0c-7ce0238dd655" (UID: "36de8e46-6ee6-4de4-bf0c-7ce0238dd655"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:01:47 crc kubenswrapper[4706]: I1206 14:01:47.579456 4706 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:47 crc kubenswrapper[4706]: I1206 14:01:47.579497 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnkkg\" (UniqueName: \"kubernetes.io/projected/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-kube-api-access-fnkkg\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:47 crc kubenswrapper[4706]: I1206 14:01:47.579509 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/36de8e46-6ee6-4de4-bf0c-7ce0238dd655-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 14:01:47 crc kubenswrapper[4706]: I1206 14:01:47.929457 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" event={"ID":"36de8e46-6ee6-4de4-bf0c-7ce0238dd655","Type":"ContainerDied","Data":"5f31b47bfc220648d39d7bfcdcdeb011b456d71d02d421c2d2e7bf944a64eb95"} Dec 06 14:01:47 crc kubenswrapper[4706]: I1206 14:01:47.929518 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f31b47bfc220648d39d7bfcdcdeb011b456d71d02d421c2d2e7bf944a64eb95" Dec 06 14:01:47 crc kubenswrapper[4706]: I1206 14:01:47.929548 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417160-7k5tx" Dec 06 14:01:58 crc kubenswrapper[4706]: I1206 14:01:58.961018 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ggc6c"] Dec 06 14:01:58 crc kubenswrapper[4706]: E1206 14:01:58.961743 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36de8e46-6ee6-4de4-bf0c-7ce0238dd655" containerName="collect-profiles" Dec 06 14:01:58 crc kubenswrapper[4706]: I1206 14:01:58.961759 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="36de8e46-6ee6-4de4-bf0c-7ce0238dd655" containerName="collect-profiles" Dec 06 14:01:58 crc kubenswrapper[4706]: I1206 14:01:58.961851 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2e4f1d6-67ff-4990-bcbf-13ad23dc511b" containerName="marketplace-operator" Dec 06 14:01:58 crc kubenswrapper[4706]: I1206 14:01:58.961860 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="36de8e46-6ee6-4de4-bf0c-7ce0238dd655" containerName="collect-profiles" Dec 06 14:01:58 crc kubenswrapper[4706]: I1206 14:01:58.962583 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ggc6c" Dec 06 14:01:58 crc kubenswrapper[4706]: I1206 14:01:58.966769 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 14:01:58 crc kubenswrapper[4706]: I1206 14:01:58.975779 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ggc6c"] Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.022470 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68mvs\" (UniqueName: \"kubernetes.io/projected/60caaf13-e871-4029-8614-dc43c10ac358-kube-api-access-68mvs\") pod \"community-operators-ggc6c\" (UID: \"60caaf13-e871-4029-8614-dc43c10ac358\") " pod="openshift-marketplace/community-operators-ggc6c" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.022553 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60caaf13-e871-4029-8614-dc43c10ac358-utilities\") pod \"community-operators-ggc6c\" (UID: \"60caaf13-e871-4029-8614-dc43c10ac358\") " pod="openshift-marketplace/community-operators-ggc6c" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.022586 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60caaf13-e871-4029-8614-dc43c10ac358-catalog-content\") pod \"community-operators-ggc6c\" (UID: \"60caaf13-e871-4029-8614-dc43c10ac358\") " pod="openshift-marketplace/community-operators-ggc6c" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.123580 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60caaf13-e871-4029-8614-dc43c10ac358-utilities\") pod \"community-operators-ggc6c\" (UID: \"60caaf13-e871-4029-8614-dc43c10ac358\") " pod="openshift-marketplace/community-operators-ggc6c" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.123633 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60caaf13-e871-4029-8614-dc43c10ac358-catalog-content\") pod \"community-operators-ggc6c\" (UID: \"60caaf13-e871-4029-8614-dc43c10ac358\") " pod="openshift-marketplace/community-operators-ggc6c" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.123717 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68mvs\" (UniqueName: \"kubernetes.io/projected/60caaf13-e871-4029-8614-dc43c10ac358-kube-api-access-68mvs\") pod \"community-operators-ggc6c\" (UID: \"60caaf13-e871-4029-8614-dc43c10ac358\") " pod="openshift-marketplace/community-operators-ggc6c" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.124187 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60caaf13-e871-4029-8614-dc43c10ac358-utilities\") pod \"community-operators-ggc6c\" (UID: \"60caaf13-e871-4029-8614-dc43c10ac358\") " pod="openshift-marketplace/community-operators-ggc6c" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.124429 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60caaf13-e871-4029-8614-dc43c10ac358-catalog-content\") pod \"community-operators-ggc6c\" (UID: \"60caaf13-e871-4029-8614-dc43c10ac358\") " pod="openshift-marketplace/community-operators-ggc6c" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.143990 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68mvs\" (UniqueName: \"kubernetes.io/projected/60caaf13-e871-4029-8614-dc43c10ac358-kube-api-access-68mvs\") pod \"community-operators-ggc6c\" (UID: \"60caaf13-e871-4029-8614-dc43c10ac358\") " pod="openshift-marketplace/community-operators-ggc6c" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.163626 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hl9mm"] Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.165007 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hl9mm" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.168104 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.177383 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hl9mm"] Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.282130 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ggc6c" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.352514 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8278b19e-f7eb-43ee-8d01-473f09b98a95-utilities\") pod \"redhat-operators-hl9mm\" (UID: \"8278b19e-f7eb-43ee-8d01-473f09b98a95\") " pod="openshift-marketplace/redhat-operators-hl9mm" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.352900 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs4vj\" (UniqueName: \"kubernetes.io/projected/8278b19e-f7eb-43ee-8d01-473f09b98a95-kube-api-access-xs4vj\") pod \"redhat-operators-hl9mm\" (UID: \"8278b19e-f7eb-43ee-8d01-473f09b98a95\") " pod="openshift-marketplace/redhat-operators-hl9mm" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.352991 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8278b19e-f7eb-43ee-8d01-473f09b98a95-catalog-content\") pod \"redhat-operators-hl9mm\" (UID: \"8278b19e-f7eb-43ee-8d01-473f09b98a95\") " pod="openshift-marketplace/redhat-operators-hl9mm" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.454677 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8278b19e-f7eb-43ee-8d01-473f09b98a95-utilities\") pod \"redhat-operators-hl9mm\" (UID: \"8278b19e-f7eb-43ee-8d01-473f09b98a95\") " pod="openshift-marketplace/redhat-operators-hl9mm" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.454745 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs4vj\" (UniqueName: \"kubernetes.io/projected/8278b19e-f7eb-43ee-8d01-473f09b98a95-kube-api-access-xs4vj\") pod \"redhat-operators-hl9mm\" (UID: \"8278b19e-f7eb-43ee-8d01-473f09b98a95\") " pod="openshift-marketplace/redhat-operators-hl9mm" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.455283 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8278b19e-f7eb-43ee-8d01-473f09b98a95-utilities\") pod \"redhat-operators-hl9mm\" (UID: \"8278b19e-f7eb-43ee-8d01-473f09b98a95\") " pod="openshift-marketplace/redhat-operators-hl9mm" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.454792 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8278b19e-f7eb-43ee-8d01-473f09b98a95-catalog-content\") pod \"redhat-operators-hl9mm\" (UID: \"8278b19e-f7eb-43ee-8d01-473f09b98a95\") " pod="openshift-marketplace/redhat-operators-hl9mm" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.455793 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8278b19e-f7eb-43ee-8d01-473f09b98a95-catalog-content\") pod \"redhat-operators-hl9mm\" (UID: \"8278b19e-f7eb-43ee-8d01-473f09b98a95\") " pod="openshift-marketplace/redhat-operators-hl9mm" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.471226 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs4vj\" (UniqueName: \"kubernetes.io/projected/8278b19e-f7eb-43ee-8d01-473f09b98a95-kube-api-access-xs4vj\") pod \"redhat-operators-hl9mm\" (UID: \"8278b19e-f7eb-43ee-8d01-473f09b98a95\") " pod="openshift-marketplace/redhat-operators-hl9mm" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.474215 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ggc6c"] Dec 06 14:01:59 crc kubenswrapper[4706]: W1206 14:01:59.477243 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60caaf13_e871_4029_8614_dc43c10ac358.slice/crio-329b52b05aebb163905639311bfcb53a8f125404ccd18f124eb6bece615fc047 WatchSource:0}: Error finding container 329b52b05aebb163905639311bfcb53a8f125404ccd18f124eb6bece615fc047: Status 404 returned error can't find the container with id 329b52b05aebb163905639311bfcb53a8f125404ccd18f124eb6bece615fc047 Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.559206 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hl9mm" Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.753376 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hl9mm"] Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.997119 4706 generic.go:334] "Generic (PLEG): container finished" podID="8278b19e-f7eb-43ee-8d01-473f09b98a95" containerID="8d6afb9dcb0f55d3d85d72ab9c11c0bfa8234bc43a7c47855b26ab1c77f14e0f" exitCode=0 Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.997200 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hl9mm" event={"ID":"8278b19e-f7eb-43ee-8d01-473f09b98a95","Type":"ContainerDied","Data":"8d6afb9dcb0f55d3d85d72ab9c11c0bfa8234bc43a7c47855b26ab1c77f14e0f"} Dec 06 14:01:59 crc kubenswrapper[4706]: I1206 14:01:59.997232 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hl9mm" event={"ID":"8278b19e-f7eb-43ee-8d01-473f09b98a95","Type":"ContainerStarted","Data":"cb4353e35fa00af4f6f4013e0d596a5250307cacaf8df064451774a73e6ceecf"} Dec 06 14:02:00 crc kubenswrapper[4706]: I1206 14:02:00.000445 4706 generic.go:334] "Generic (PLEG): container finished" podID="60caaf13-e871-4029-8614-dc43c10ac358" containerID="3dc3b99c8105bdb6973f1ec1e713d6cbee16b7a50dc8c60fffd501dac0a57427" exitCode=0 Dec 06 14:02:00 crc kubenswrapper[4706]: I1206 14:02:00.000470 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggc6c" event={"ID":"60caaf13-e871-4029-8614-dc43c10ac358","Type":"ContainerDied","Data":"3dc3b99c8105bdb6973f1ec1e713d6cbee16b7a50dc8c60fffd501dac0a57427"} Dec 06 14:02:00 crc kubenswrapper[4706]: I1206 14:02:00.000492 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggc6c" event={"ID":"60caaf13-e871-4029-8614-dc43c10ac358","Type":"ContainerStarted","Data":"329b52b05aebb163905639311bfcb53a8f125404ccd18f124eb6bece615fc047"} Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.020729 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggc6c" event={"ID":"60caaf13-e871-4029-8614-dc43c10ac358","Type":"ContainerStarted","Data":"c69bfe28cb4e94248b70296c7657ddaa41fe2d9787f7b6768df31a59277e1ef7"} Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.360553 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ssx77"] Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.361479 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ssx77" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.364066 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.371913 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ssx77"] Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.487487 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110-utilities\") pod \"certified-operators-ssx77\" (UID: \"6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110\") " pod="openshift-marketplace/certified-operators-ssx77" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.487542 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110-catalog-content\") pod \"certified-operators-ssx77\" (UID: \"6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110\") " pod="openshift-marketplace/certified-operators-ssx77" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.487573 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nnd5\" (UniqueName: \"kubernetes.io/projected/6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110-kube-api-access-4nnd5\") pod \"certified-operators-ssx77\" (UID: \"6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110\") " pod="openshift-marketplace/certified-operators-ssx77" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.556738 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-62dkd"] Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.557744 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-62dkd" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.562233 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.569562 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-62dkd"] Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.588903 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0514bb3a-f280-4208-9dff-ca5549dff0ca-catalog-content\") pod \"redhat-marketplace-62dkd\" (UID: \"0514bb3a-f280-4208-9dff-ca5549dff0ca\") " pod="openshift-marketplace/redhat-marketplace-62dkd" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.589151 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn4cz\" (UniqueName: \"kubernetes.io/projected/0514bb3a-f280-4208-9dff-ca5549dff0ca-kube-api-access-fn4cz\") pod \"redhat-marketplace-62dkd\" (UID: \"0514bb3a-f280-4208-9dff-ca5549dff0ca\") " pod="openshift-marketplace/redhat-marketplace-62dkd" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.590367 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110-utilities\") pod \"certified-operators-ssx77\" (UID: \"6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110\") " pod="openshift-marketplace/certified-operators-ssx77" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.590445 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110-catalog-content\") pod \"certified-operators-ssx77\" (UID: \"6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110\") " pod="openshift-marketplace/certified-operators-ssx77" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.590515 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0514bb3a-f280-4208-9dff-ca5549dff0ca-utilities\") pod \"redhat-marketplace-62dkd\" (UID: \"0514bb3a-f280-4208-9dff-ca5549dff0ca\") " pod="openshift-marketplace/redhat-marketplace-62dkd" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.590554 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nnd5\" (UniqueName: \"kubernetes.io/projected/6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110-kube-api-access-4nnd5\") pod \"certified-operators-ssx77\" (UID: \"6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110\") " pod="openshift-marketplace/certified-operators-ssx77" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.592535 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110-utilities\") pod \"certified-operators-ssx77\" (UID: \"6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110\") " pod="openshift-marketplace/certified-operators-ssx77" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.593024 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110-catalog-content\") pod \"certified-operators-ssx77\" (UID: \"6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110\") " pod="openshift-marketplace/certified-operators-ssx77" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.621173 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nnd5\" (UniqueName: \"kubernetes.io/projected/6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110-kube-api-access-4nnd5\") pod \"certified-operators-ssx77\" (UID: \"6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110\") " pod="openshift-marketplace/certified-operators-ssx77" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.691353 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn4cz\" (UniqueName: \"kubernetes.io/projected/0514bb3a-f280-4208-9dff-ca5549dff0ca-kube-api-access-fn4cz\") pod \"redhat-marketplace-62dkd\" (UID: \"0514bb3a-f280-4208-9dff-ca5549dff0ca\") " pod="openshift-marketplace/redhat-marketplace-62dkd" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.691462 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0514bb3a-f280-4208-9dff-ca5549dff0ca-utilities\") pod \"redhat-marketplace-62dkd\" (UID: \"0514bb3a-f280-4208-9dff-ca5549dff0ca\") " pod="openshift-marketplace/redhat-marketplace-62dkd" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.691516 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0514bb3a-f280-4208-9dff-ca5549dff0ca-catalog-content\") pod \"redhat-marketplace-62dkd\" (UID: \"0514bb3a-f280-4208-9dff-ca5549dff0ca\") " pod="openshift-marketplace/redhat-marketplace-62dkd" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.692158 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0514bb3a-f280-4208-9dff-ca5549dff0ca-catalog-content\") pod \"redhat-marketplace-62dkd\" (UID: \"0514bb3a-f280-4208-9dff-ca5549dff0ca\") " pod="openshift-marketplace/redhat-marketplace-62dkd" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.692669 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0514bb3a-f280-4208-9dff-ca5549dff0ca-utilities\") pod \"redhat-marketplace-62dkd\" (UID: \"0514bb3a-f280-4208-9dff-ca5549dff0ca\") " pod="openshift-marketplace/redhat-marketplace-62dkd" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.713285 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn4cz\" (UniqueName: \"kubernetes.io/projected/0514bb3a-f280-4208-9dff-ca5549dff0ca-kube-api-access-fn4cz\") pod \"redhat-marketplace-62dkd\" (UID: \"0514bb3a-f280-4208-9dff-ca5549dff0ca\") " pod="openshift-marketplace/redhat-marketplace-62dkd" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.718711 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ssx77" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.878509 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-62dkd" Dec 06 14:02:01 crc kubenswrapper[4706]: I1206 14:02:01.941033 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ssx77"] Dec 06 14:02:01 crc kubenswrapper[4706]: W1206 14:02:01.947360 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ea6c2f3_7a03_4e0a_9be9_be2a62a5f110.slice/crio-afd27b09e6ebd1f9d31b315359b24894605fea4f297a6fd75fa39ed3b3d581c3 WatchSource:0}: Error finding container afd27b09e6ebd1f9d31b315359b24894605fea4f297a6fd75fa39ed3b3d581c3: Status 404 returned error can't find the container with id afd27b09e6ebd1f9d31b315359b24894605fea4f297a6fd75fa39ed3b3d581c3 Dec 06 14:02:02 crc kubenswrapper[4706]: I1206 14:02:02.033244 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ssx77" event={"ID":"6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110","Type":"ContainerStarted","Data":"afd27b09e6ebd1f9d31b315359b24894605fea4f297a6fd75fa39ed3b3d581c3"} Dec 06 14:02:02 crc kubenswrapper[4706]: I1206 14:02:02.037514 4706 generic.go:334] "Generic (PLEG): container finished" podID="8278b19e-f7eb-43ee-8d01-473f09b98a95" containerID="d2bf6b96cfbef1484d2472038456eb12ffd64e6a65a0bd461107dba244bb5395" exitCode=0 Dec 06 14:02:02 crc kubenswrapper[4706]: I1206 14:02:02.037579 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hl9mm" event={"ID":"8278b19e-f7eb-43ee-8d01-473f09b98a95","Type":"ContainerDied","Data":"d2bf6b96cfbef1484d2472038456eb12ffd64e6a65a0bd461107dba244bb5395"} Dec 06 14:02:02 crc kubenswrapper[4706]: I1206 14:02:02.045587 4706 generic.go:334] "Generic (PLEG): container finished" podID="60caaf13-e871-4029-8614-dc43c10ac358" containerID="c69bfe28cb4e94248b70296c7657ddaa41fe2d9787f7b6768df31a59277e1ef7" exitCode=0 Dec 06 14:02:02 crc kubenswrapper[4706]: I1206 14:02:02.045630 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggc6c" event={"ID":"60caaf13-e871-4029-8614-dc43c10ac358","Type":"ContainerDied","Data":"c69bfe28cb4e94248b70296c7657ddaa41fe2d9787f7b6768df31a59277e1ef7"} Dec 06 14:02:02 crc kubenswrapper[4706]: I1206 14:02:02.113565 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-62dkd"] Dec 06 14:02:02 crc kubenswrapper[4706]: W1206 14:02:02.115611 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0514bb3a_f280_4208_9dff_ca5549dff0ca.slice/crio-4e1836b7fcc7332cd5088e9e892476324b9310fc4013d8e75b4080456355cf66 WatchSource:0}: Error finding container 4e1836b7fcc7332cd5088e9e892476324b9310fc4013d8e75b4080456355cf66: Status 404 returned error can't find the container with id 4e1836b7fcc7332cd5088e9e892476324b9310fc4013d8e75b4080456355cf66 Dec 06 14:02:03 crc kubenswrapper[4706]: I1206 14:02:03.056104 4706 generic.go:334] "Generic (PLEG): container finished" podID="6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110" containerID="41f47f216c0d6ed6f35bc5032e9df2357a54260ed848c19a26b4156074d32419" exitCode=0 Dec 06 14:02:03 crc kubenswrapper[4706]: I1206 14:02:03.056214 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ssx77" event={"ID":"6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110","Type":"ContainerDied","Data":"41f47f216c0d6ed6f35bc5032e9df2357a54260ed848c19a26b4156074d32419"} Dec 06 14:02:03 crc kubenswrapper[4706]: I1206 14:02:03.062022 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hl9mm" event={"ID":"8278b19e-f7eb-43ee-8d01-473f09b98a95","Type":"ContainerStarted","Data":"6c2e0c5d56d5f07b07a6c0e977811be2deeeb0e0c169f5a42d6f75c6f6cd68d6"} Dec 06 14:02:03 crc kubenswrapper[4706]: I1206 14:02:03.065265 4706 generic.go:334] "Generic (PLEG): container finished" podID="0514bb3a-f280-4208-9dff-ca5549dff0ca" containerID="976fe7406c3c76edc2fabeee08be0499f730b21cd283516b90754ab72b006d90" exitCode=0 Dec 06 14:02:03 crc kubenswrapper[4706]: I1206 14:02:03.066142 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62dkd" event={"ID":"0514bb3a-f280-4208-9dff-ca5549dff0ca","Type":"ContainerDied","Data":"976fe7406c3c76edc2fabeee08be0499f730b21cd283516b90754ab72b006d90"} Dec 06 14:02:03 crc kubenswrapper[4706]: I1206 14:02:03.066186 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62dkd" event={"ID":"0514bb3a-f280-4208-9dff-ca5549dff0ca","Type":"ContainerStarted","Data":"4e1836b7fcc7332cd5088e9e892476324b9310fc4013d8e75b4080456355cf66"} Dec 06 14:02:03 crc kubenswrapper[4706]: I1206 14:02:03.072044 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggc6c" event={"ID":"60caaf13-e871-4029-8614-dc43c10ac358","Type":"ContainerStarted","Data":"1110ec26e92d9a2e47a05607d2ada556d45c16889cfca8622e84a954ba0e4681"} Dec 06 14:02:03 crc kubenswrapper[4706]: I1206 14:02:03.128181 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hl9mm" podStartSLOduration=1.5517549019999999 podStartE2EDuration="4.128153722s" podCreationTimestamp="2025-12-06 14:01:59 +0000 UTC" firstStartedPulling="2025-12-06 14:01:59.998579444 +0000 UTC m=+362.834372311" lastFinishedPulling="2025-12-06 14:02:02.574978274 +0000 UTC m=+365.410771131" observedRunningTime="2025-12-06 14:02:03.121181106 +0000 UTC m=+365.956973973" watchObservedRunningTime="2025-12-06 14:02:03.128153722 +0000 UTC m=+365.963946619" Dec 06 14:02:03 crc kubenswrapper[4706]: I1206 14:02:03.150777 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ggc6c" podStartSLOduration=2.6794783349999998 podStartE2EDuration="5.150752906s" podCreationTimestamp="2025-12-06 14:01:58 +0000 UTC" firstStartedPulling="2025-12-06 14:02:00.001872612 +0000 UTC m=+362.837665479" lastFinishedPulling="2025-12-06 14:02:02.473147193 +0000 UTC m=+365.308940050" observedRunningTime="2025-12-06 14:02:03.14752795 +0000 UTC m=+365.983320847" watchObservedRunningTime="2025-12-06 14:02:03.150752906 +0000 UTC m=+365.986545793" Dec 06 14:02:04 crc kubenswrapper[4706]: I1206 14:02:04.080074 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ssx77" event={"ID":"6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110","Type":"ContainerStarted","Data":"e8edcf2a3c1703f1638436a52869fde2317bda58ab2ef385ad971ab32fb9f9e7"} Dec 06 14:02:04 crc kubenswrapper[4706]: I1206 14:02:04.082205 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62dkd" event={"ID":"0514bb3a-f280-4208-9dff-ca5549dff0ca","Type":"ContainerStarted","Data":"fbd571889455b270fad821d3fca77d614dcc67a66eaadb69bf05bfd2981ab84d"} Dec 06 14:02:05 crc kubenswrapper[4706]: I1206 14:02:05.093470 4706 generic.go:334] "Generic (PLEG): container finished" podID="6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110" containerID="e8edcf2a3c1703f1638436a52869fde2317bda58ab2ef385ad971ab32fb9f9e7" exitCode=0 Dec 06 14:02:05 crc kubenswrapper[4706]: I1206 14:02:05.093615 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ssx77" event={"ID":"6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110","Type":"ContainerDied","Data":"e8edcf2a3c1703f1638436a52869fde2317bda58ab2ef385ad971ab32fb9f9e7"} Dec 06 14:02:05 crc kubenswrapper[4706]: I1206 14:02:05.098412 4706 generic.go:334] "Generic (PLEG): container finished" podID="0514bb3a-f280-4208-9dff-ca5549dff0ca" containerID="fbd571889455b270fad821d3fca77d614dcc67a66eaadb69bf05bfd2981ab84d" exitCode=0 Dec 06 14:02:05 crc kubenswrapper[4706]: I1206 14:02:05.098462 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62dkd" event={"ID":"0514bb3a-f280-4208-9dff-ca5549dff0ca","Type":"ContainerDied","Data":"fbd571889455b270fad821d3fca77d614dcc67a66eaadb69bf05bfd2981ab84d"} Dec 06 14:02:07 crc kubenswrapper[4706]: I1206 14:02:07.111529 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62dkd" event={"ID":"0514bb3a-f280-4208-9dff-ca5549dff0ca","Type":"ContainerStarted","Data":"b143b3dcfade0b2ad0b21d5b92474b7530bbaed4cb976d0a772d971028ffef55"} Dec 06 14:02:07 crc kubenswrapper[4706]: I1206 14:02:07.115240 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ssx77" event={"ID":"6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110","Type":"ContainerStarted","Data":"258c73769e0326737f82a0691bd2f997ad8fb857e069515e40b93333c28a8683"} Dec 06 14:02:07 crc kubenswrapper[4706]: I1206 14:02:07.135499 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-62dkd" podStartSLOduration=3.401286645 podStartE2EDuration="6.13548264s" podCreationTimestamp="2025-12-06 14:02:01 +0000 UTC" firstStartedPulling="2025-12-06 14:02:03.067496512 +0000 UTC m=+365.903289379" lastFinishedPulling="2025-12-06 14:02:05.801692507 +0000 UTC m=+368.637485374" observedRunningTime="2025-12-06 14:02:07.132191742 +0000 UTC m=+369.967984609" watchObservedRunningTime="2025-12-06 14:02:07.13548264 +0000 UTC m=+369.971275497" Dec 06 14:02:07 crc kubenswrapper[4706]: I1206 14:02:07.157070 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ssx77" podStartSLOduration=3.460200328 podStartE2EDuration="6.157047116s" podCreationTimestamp="2025-12-06 14:02:01 +0000 UTC" firstStartedPulling="2025-12-06 14:02:03.058609184 +0000 UTC m=+365.894402051" lastFinishedPulling="2025-12-06 14:02:05.755455982 +0000 UTC m=+368.591248839" observedRunningTime="2025-12-06 14:02:07.152911516 +0000 UTC m=+369.988704373" watchObservedRunningTime="2025-12-06 14:02:07.157047116 +0000 UTC m=+369.992839973" Dec 06 14:02:09 crc kubenswrapper[4706]: I1206 14:02:09.283013 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ggc6c" Dec 06 14:02:09 crc kubenswrapper[4706]: I1206 14:02:09.283111 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ggc6c" Dec 06 14:02:09 crc kubenswrapper[4706]: I1206 14:02:09.341353 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ggc6c" Dec 06 14:02:09 crc kubenswrapper[4706]: I1206 14:02:09.559940 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hl9mm" Dec 06 14:02:09 crc kubenswrapper[4706]: I1206 14:02:09.560018 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hl9mm" Dec 06 14:02:09 crc kubenswrapper[4706]: I1206 14:02:09.622499 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hl9mm" Dec 06 14:02:09 crc kubenswrapper[4706]: I1206 14:02:09.895180 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tsvsk"] Dec 06 14:02:09 crc kubenswrapper[4706]: I1206 14:02:09.895725 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" podUID="475504b3-7830-4226-b00c-3d35eda647fe" containerName="controller-manager" containerID="cri-o://84d8095aad61a3e7266019b5357582efdd2a1549358ef81da95f9ea47ecf7cee" gracePeriod=30 Dec 06 14:02:10 crc kubenswrapper[4706]: I1206 14:02:10.004749 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8"] Dec 06 14:02:10 crc kubenswrapper[4706]: I1206 14:02:10.005169 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" podUID="b801f158-2df3-44c7-8bfe-5caa8467e29d" containerName="route-controller-manager" containerID="cri-o://41e759f55161abf22d950b30fa767aeed37879f3431c5aacf9f23cbdc1040c2a" gracePeriod=30 Dec 06 14:02:10 crc kubenswrapper[4706]: I1206 14:02:10.177095 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hl9mm" Dec 06 14:02:10 crc kubenswrapper[4706]: I1206 14:02:10.193150 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ggc6c" Dec 06 14:02:10 crc kubenswrapper[4706]: I1206 14:02:10.549245 4706 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-tsvsk container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 06 14:02:10 crc kubenswrapper[4706]: I1206 14:02:10.549395 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" podUID="475504b3-7830-4226-b00c-3d35eda647fe" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 06 14:02:11 crc kubenswrapper[4706]: I1206 14:02:11.719457 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ssx77" Dec 06 14:02:11 crc kubenswrapper[4706]: I1206 14:02:11.719686 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ssx77" Dec 06 14:02:11 crc kubenswrapper[4706]: I1206 14:02:11.772091 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ssx77" Dec 06 14:02:11 crc kubenswrapper[4706]: I1206 14:02:11.847246 4706 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-dmfw8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 06 14:02:11 crc kubenswrapper[4706]: I1206 14:02:11.847332 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" podUID="b801f158-2df3-44c7-8bfe-5caa8467e29d" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 06 14:02:11 crc kubenswrapper[4706]: I1206 14:02:11.879917 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-62dkd" Dec 06 14:02:11 crc kubenswrapper[4706]: I1206 14:02:11.879986 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-62dkd" Dec 06 14:02:11 crc kubenswrapper[4706]: I1206 14:02:11.939336 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-62dkd" Dec 06 14:02:12 crc kubenswrapper[4706]: I1206 14:02:12.192601 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ssx77" Dec 06 14:02:12 crc kubenswrapper[4706]: I1206 14:02:12.206067 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-62dkd" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.006243 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.035813 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz"] Dec 06 14:02:13 crc kubenswrapper[4706]: E1206 14:02:13.036049 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="475504b3-7830-4226-b00c-3d35eda647fe" containerName="controller-manager" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.036063 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="475504b3-7830-4226-b00c-3d35eda647fe" containerName="controller-manager" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.036182 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="475504b3-7830-4226-b00c-3d35eda647fe" containerName="controller-manager" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.036630 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.050095 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz"] Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.059218 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.142633 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-client-ca\") pod \"475504b3-7830-4226-b00c-3d35eda647fe\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.142679 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8xz9\" (UniqueName: \"kubernetes.io/projected/b801f158-2df3-44c7-8bfe-5caa8467e29d-kube-api-access-d8xz9\") pod \"b801f158-2df3-44c7-8bfe-5caa8467e29d\" (UID: \"b801f158-2df3-44c7-8bfe-5caa8467e29d\") " Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.142701 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b801f158-2df3-44c7-8bfe-5caa8467e29d-client-ca\") pod \"b801f158-2df3-44c7-8bfe-5caa8467e29d\" (UID: \"b801f158-2df3-44c7-8bfe-5caa8467e29d\") " Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.142722 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8k5g\" (UniqueName: \"kubernetes.io/projected/475504b3-7830-4226-b00c-3d35eda647fe-kube-api-access-p8k5g\") pod \"475504b3-7830-4226-b00c-3d35eda647fe\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.142759 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b801f158-2df3-44c7-8bfe-5caa8467e29d-serving-cert\") pod \"b801f158-2df3-44c7-8bfe-5caa8467e29d\" (UID: \"b801f158-2df3-44c7-8bfe-5caa8467e29d\") " Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.142784 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/475504b3-7830-4226-b00c-3d35eda647fe-serving-cert\") pod \"475504b3-7830-4226-b00c-3d35eda647fe\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.142803 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-proxy-ca-bundles\") pod \"475504b3-7830-4226-b00c-3d35eda647fe\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.142863 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b801f158-2df3-44c7-8bfe-5caa8467e29d-config\") pod \"b801f158-2df3-44c7-8bfe-5caa8467e29d\" (UID: \"b801f158-2df3-44c7-8bfe-5caa8467e29d\") " Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.142887 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-config\") pod \"475504b3-7830-4226-b00c-3d35eda647fe\" (UID: \"475504b3-7830-4226-b00c-3d35eda647fe\") " Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.143048 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89efb534-2dd5-40bf-b4f0-538f2c3c58f3-config\") pod \"controller-manager-67f9b8c6bd-smkkz\" (UID: \"89efb534-2dd5-40bf-b4f0-538f2c3c58f3\") " pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.143076 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdrnr\" (UniqueName: \"kubernetes.io/projected/89efb534-2dd5-40bf-b4f0-538f2c3c58f3-kube-api-access-rdrnr\") pod \"controller-manager-67f9b8c6bd-smkkz\" (UID: \"89efb534-2dd5-40bf-b4f0-538f2c3c58f3\") " pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.143098 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89efb534-2dd5-40bf-b4f0-538f2c3c58f3-client-ca\") pod \"controller-manager-67f9b8c6bd-smkkz\" (UID: \"89efb534-2dd5-40bf-b4f0-538f2c3c58f3\") " pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.143125 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89efb534-2dd5-40bf-b4f0-538f2c3c58f3-serving-cert\") pod \"controller-manager-67f9b8c6bd-smkkz\" (UID: \"89efb534-2dd5-40bf-b4f0-538f2c3c58f3\") " pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.143156 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/89efb534-2dd5-40bf-b4f0-538f2c3c58f3-proxy-ca-bundles\") pod \"controller-manager-67f9b8c6bd-smkkz\" (UID: \"89efb534-2dd5-40bf-b4f0-538f2c3c58f3\") " pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.143623 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b801f158-2df3-44c7-8bfe-5caa8467e29d-client-ca" (OuterVolumeSpecName: "client-ca") pod "b801f158-2df3-44c7-8bfe-5caa8467e29d" (UID: "b801f158-2df3-44c7-8bfe-5caa8467e29d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.143651 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b801f158-2df3-44c7-8bfe-5caa8467e29d-config" (OuterVolumeSpecName: "config") pod "b801f158-2df3-44c7-8bfe-5caa8467e29d" (UID: "b801f158-2df3-44c7-8bfe-5caa8467e29d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.143666 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "475504b3-7830-4226-b00c-3d35eda647fe" (UID: "475504b3-7830-4226-b00c-3d35eda647fe"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.143726 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-config" (OuterVolumeSpecName: "config") pod "475504b3-7830-4226-b00c-3d35eda647fe" (UID: "475504b3-7830-4226-b00c-3d35eda647fe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.144107 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-client-ca" (OuterVolumeSpecName: "client-ca") pod "475504b3-7830-4226-b00c-3d35eda647fe" (UID: "475504b3-7830-4226-b00c-3d35eda647fe"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.148007 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/475504b3-7830-4226-b00c-3d35eda647fe-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "475504b3-7830-4226-b00c-3d35eda647fe" (UID: "475504b3-7830-4226-b00c-3d35eda647fe"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.148948 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/475504b3-7830-4226-b00c-3d35eda647fe-kube-api-access-p8k5g" (OuterVolumeSpecName: "kube-api-access-p8k5g") pod "475504b3-7830-4226-b00c-3d35eda647fe" (UID: "475504b3-7830-4226-b00c-3d35eda647fe"). InnerVolumeSpecName "kube-api-access-p8k5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.153562 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b801f158-2df3-44c7-8bfe-5caa8467e29d-kube-api-access-d8xz9" (OuterVolumeSpecName: "kube-api-access-d8xz9") pod "b801f158-2df3-44c7-8bfe-5caa8467e29d" (UID: "b801f158-2df3-44c7-8bfe-5caa8467e29d"). InnerVolumeSpecName "kube-api-access-d8xz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.154824 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b801f158-2df3-44c7-8bfe-5caa8467e29d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b801f158-2df3-44c7-8bfe-5caa8467e29d" (UID: "b801f158-2df3-44c7-8bfe-5caa8467e29d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.157175 4706 generic.go:334] "Generic (PLEG): container finished" podID="b801f158-2df3-44c7-8bfe-5caa8467e29d" containerID="41e759f55161abf22d950b30fa767aeed37879f3431c5aacf9f23cbdc1040c2a" exitCode=0 Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.157241 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" event={"ID":"b801f158-2df3-44c7-8bfe-5caa8467e29d","Type":"ContainerDied","Data":"41e759f55161abf22d950b30fa767aeed37879f3431c5aacf9f23cbdc1040c2a"} Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.157271 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" event={"ID":"b801f158-2df3-44c7-8bfe-5caa8467e29d","Type":"ContainerDied","Data":"a4c30054331e61c6b1cb168dbe02d6eedccf1da77ed92b65306384af25abbddd"} Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.157289 4706 scope.go:117] "RemoveContainer" containerID="41e759f55161abf22d950b30fa767aeed37879f3431c5aacf9f23cbdc1040c2a" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.157244 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.163493 4706 generic.go:334] "Generic (PLEG): container finished" podID="475504b3-7830-4226-b00c-3d35eda647fe" containerID="84d8095aad61a3e7266019b5357582efdd2a1549358ef81da95f9ea47ecf7cee" exitCode=0 Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.163577 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.163640 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" event={"ID":"475504b3-7830-4226-b00c-3d35eda647fe","Type":"ContainerDied","Data":"84d8095aad61a3e7266019b5357582efdd2a1549358ef81da95f9ea47ecf7cee"} Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.163671 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tsvsk" event={"ID":"475504b3-7830-4226-b00c-3d35eda647fe","Type":"ContainerDied","Data":"b0aaca3181268ca5d05f5decc481795b9f531d9bd6ebe9f9384b93ea16a3f08a"} Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.180618 4706 scope.go:117] "RemoveContainer" containerID="41e759f55161abf22d950b30fa767aeed37879f3431c5aacf9f23cbdc1040c2a" Dec 06 14:02:13 crc kubenswrapper[4706]: E1206 14:02:13.181908 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41e759f55161abf22d950b30fa767aeed37879f3431c5aacf9f23cbdc1040c2a\": container with ID starting with 41e759f55161abf22d950b30fa767aeed37879f3431c5aacf9f23cbdc1040c2a not found: ID does not exist" containerID="41e759f55161abf22d950b30fa767aeed37879f3431c5aacf9f23cbdc1040c2a" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.181949 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41e759f55161abf22d950b30fa767aeed37879f3431c5aacf9f23cbdc1040c2a"} err="failed to get container status \"41e759f55161abf22d950b30fa767aeed37879f3431c5aacf9f23cbdc1040c2a\": rpc error: code = NotFound desc = could not find container \"41e759f55161abf22d950b30fa767aeed37879f3431c5aacf9f23cbdc1040c2a\": container with ID starting with 41e759f55161abf22d950b30fa767aeed37879f3431c5aacf9f23cbdc1040c2a not found: ID does not exist" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.181976 4706 scope.go:117] "RemoveContainer" containerID="84d8095aad61a3e7266019b5357582efdd2a1549358ef81da95f9ea47ecf7cee" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.193643 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8"] Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.197909 4706 scope.go:117] "RemoveContainer" containerID="84d8095aad61a3e7266019b5357582efdd2a1549358ef81da95f9ea47ecf7cee" Dec 06 14:02:13 crc kubenswrapper[4706]: E1206 14:02:13.198252 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84d8095aad61a3e7266019b5357582efdd2a1549358ef81da95f9ea47ecf7cee\": container with ID starting with 84d8095aad61a3e7266019b5357582efdd2a1549358ef81da95f9ea47ecf7cee not found: ID does not exist" containerID="84d8095aad61a3e7266019b5357582efdd2a1549358ef81da95f9ea47ecf7cee" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.198281 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84d8095aad61a3e7266019b5357582efdd2a1549358ef81da95f9ea47ecf7cee"} err="failed to get container status \"84d8095aad61a3e7266019b5357582efdd2a1549358ef81da95f9ea47ecf7cee\": rpc error: code = NotFound desc = could not find container \"84d8095aad61a3e7266019b5357582efdd2a1549358ef81da95f9ea47ecf7cee\": container with ID starting with 84d8095aad61a3e7266019b5357582efdd2a1549358ef81da95f9ea47ecf7cee not found: ID does not exist" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.203298 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dmfw8"] Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.208169 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tsvsk"] Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.213075 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tsvsk"] Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.244674 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89efb534-2dd5-40bf-b4f0-538f2c3c58f3-client-ca\") pod \"controller-manager-67f9b8c6bd-smkkz\" (UID: \"89efb534-2dd5-40bf-b4f0-538f2c3c58f3\") " pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.244738 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89efb534-2dd5-40bf-b4f0-538f2c3c58f3-serving-cert\") pod \"controller-manager-67f9b8c6bd-smkkz\" (UID: \"89efb534-2dd5-40bf-b4f0-538f2c3c58f3\") " pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.244799 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/89efb534-2dd5-40bf-b4f0-538f2c3c58f3-proxy-ca-bundles\") pod \"controller-manager-67f9b8c6bd-smkkz\" (UID: \"89efb534-2dd5-40bf-b4f0-538f2c3c58f3\") " pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.244871 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89efb534-2dd5-40bf-b4f0-538f2c3c58f3-config\") pod \"controller-manager-67f9b8c6bd-smkkz\" (UID: \"89efb534-2dd5-40bf-b4f0-538f2c3c58f3\") " pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.244903 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdrnr\" (UniqueName: \"kubernetes.io/projected/89efb534-2dd5-40bf-b4f0-538f2c3c58f3-kube-api-access-rdrnr\") pod \"controller-manager-67f9b8c6bd-smkkz\" (UID: \"89efb534-2dd5-40bf-b4f0-538f2c3c58f3\") " pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.244946 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b801f158-2df3-44c7-8bfe-5caa8467e29d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.244961 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/475504b3-7830-4226-b00c-3d35eda647fe-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.244974 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.244989 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b801f158-2df3-44c7-8bfe-5caa8467e29d-config\") on node \"crc\" DevicePath \"\"" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.245003 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-config\") on node \"crc\" DevicePath \"\"" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.245015 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/475504b3-7830-4226-b00c-3d35eda647fe-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.245027 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8xz9\" (UniqueName: \"kubernetes.io/projected/b801f158-2df3-44c7-8bfe-5caa8467e29d-kube-api-access-d8xz9\") on node \"crc\" DevicePath \"\"" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.245038 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b801f158-2df3-44c7-8bfe-5caa8467e29d-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.245051 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8k5g\" (UniqueName: \"kubernetes.io/projected/475504b3-7830-4226-b00c-3d35eda647fe-kube-api-access-p8k5g\") on node \"crc\" DevicePath \"\"" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.246247 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89efb534-2dd5-40bf-b4f0-538f2c3c58f3-client-ca\") pod \"controller-manager-67f9b8c6bd-smkkz\" (UID: \"89efb534-2dd5-40bf-b4f0-538f2c3c58f3\") " pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.246651 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/89efb534-2dd5-40bf-b4f0-538f2c3c58f3-proxy-ca-bundles\") pod \"controller-manager-67f9b8c6bd-smkkz\" (UID: \"89efb534-2dd5-40bf-b4f0-538f2c3c58f3\") " pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.248709 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89efb534-2dd5-40bf-b4f0-538f2c3c58f3-config\") pod \"controller-manager-67f9b8c6bd-smkkz\" (UID: \"89efb534-2dd5-40bf-b4f0-538f2c3c58f3\") " pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.248927 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89efb534-2dd5-40bf-b4f0-538f2c3c58f3-serving-cert\") pod \"controller-manager-67f9b8c6bd-smkkz\" (UID: \"89efb534-2dd5-40bf-b4f0-538f2c3c58f3\") " pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.274000 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdrnr\" (UniqueName: \"kubernetes.io/projected/89efb534-2dd5-40bf-b4f0-538f2c3c58f3-kube-api-access-rdrnr\") pod \"controller-manager-67f9b8c6bd-smkkz\" (UID: \"89efb534-2dd5-40bf-b4f0-538f2c3c58f3\") " pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.370430 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.537400 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="475504b3-7830-4226-b00c-3d35eda647fe" path="/var/lib/kubelet/pods/475504b3-7830-4226-b00c-3d35eda647fe/volumes" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.539075 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b801f158-2df3-44c7-8bfe-5caa8467e29d" path="/var/lib/kubelet/pods/b801f158-2df3-44c7-8bfe-5caa8467e29d/volumes" Dec 06 14:02:13 crc kubenswrapper[4706]: I1206 14:02:13.643146 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz"] Dec 06 14:02:14 crc kubenswrapper[4706]: I1206 14:02:14.176505 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" event={"ID":"89efb534-2dd5-40bf-b4f0-538f2c3c58f3","Type":"ContainerStarted","Data":"67d199d25e39cab708a0b85f090573e35c35aa543381fecfb263cb4659278c1a"} Dec 06 14:02:14 crc kubenswrapper[4706]: I1206 14:02:14.580670 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:02:14 crc kubenswrapper[4706]: I1206 14:02:14.580756 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:02:15 crc kubenswrapper[4706]: I1206 14:02:15.188137 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" event={"ID":"89efb534-2dd5-40bf-b4f0-538f2c3c58f3","Type":"ContainerStarted","Data":"77553d3688e0bd90905b592c9c392e76394d4da94dd958502361df59e19503a4"} Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.020190 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62"] Dec 06 14:02:16 crc kubenswrapper[4706]: E1206 14:02:16.020662 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b801f158-2df3-44c7-8bfe-5caa8467e29d" containerName="route-controller-manager" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.020743 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b801f158-2df3-44c7-8bfe-5caa8467e29d" containerName="route-controller-manager" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.020902 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b801f158-2df3-44c7-8bfe-5caa8467e29d" containerName="route-controller-manager" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.021430 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.025195 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.025903 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.026411 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.036898 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.037096 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.037222 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.054119 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62"] Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.083300 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7aa3a14-1317-4a45-945b-4bb6eb26c9ea-client-ca\") pod \"route-controller-manager-94f88b9db-vth62\" (UID: \"b7aa3a14-1317-4a45-945b-4bb6eb26c9ea\") " pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.083417 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7aa3a14-1317-4a45-945b-4bb6eb26c9ea-config\") pod \"route-controller-manager-94f88b9db-vth62\" (UID: \"b7aa3a14-1317-4a45-945b-4bb6eb26c9ea\") " pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.083457 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7aa3a14-1317-4a45-945b-4bb6eb26c9ea-serving-cert\") pod \"route-controller-manager-94f88b9db-vth62\" (UID: \"b7aa3a14-1317-4a45-945b-4bb6eb26c9ea\") " pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.083532 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9mfh\" (UniqueName: \"kubernetes.io/projected/b7aa3a14-1317-4a45-945b-4bb6eb26c9ea-kube-api-access-m9mfh\") pod \"route-controller-manager-94f88b9db-vth62\" (UID: \"b7aa3a14-1317-4a45-945b-4bb6eb26c9ea\") " pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.185121 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7aa3a14-1317-4a45-945b-4bb6eb26c9ea-config\") pod \"route-controller-manager-94f88b9db-vth62\" (UID: \"b7aa3a14-1317-4a45-945b-4bb6eb26c9ea\") " pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.185204 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7aa3a14-1317-4a45-945b-4bb6eb26c9ea-client-ca\") pod \"route-controller-manager-94f88b9db-vth62\" (UID: \"b7aa3a14-1317-4a45-945b-4bb6eb26c9ea\") " pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.185352 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7aa3a14-1317-4a45-945b-4bb6eb26c9ea-serving-cert\") pod \"route-controller-manager-94f88b9db-vth62\" (UID: \"b7aa3a14-1317-4a45-945b-4bb6eb26c9ea\") " pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.185435 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9mfh\" (UniqueName: \"kubernetes.io/projected/b7aa3a14-1317-4a45-945b-4bb6eb26c9ea-kube-api-access-m9mfh\") pod \"route-controller-manager-94f88b9db-vth62\" (UID: \"b7aa3a14-1317-4a45-945b-4bb6eb26c9ea\") " pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.186712 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7aa3a14-1317-4a45-945b-4bb6eb26c9ea-client-ca\") pod \"route-controller-manager-94f88b9db-vth62\" (UID: \"b7aa3a14-1317-4a45-945b-4bb6eb26c9ea\") " pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.187197 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7aa3a14-1317-4a45-945b-4bb6eb26c9ea-config\") pod \"route-controller-manager-94f88b9db-vth62\" (UID: \"b7aa3a14-1317-4a45-945b-4bb6eb26c9ea\") " pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.191917 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7aa3a14-1317-4a45-945b-4bb6eb26c9ea-serving-cert\") pod \"route-controller-manager-94f88b9db-vth62\" (UID: \"b7aa3a14-1317-4a45-945b-4bb6eb26c9ea\") " pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.195430 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.202489 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.211163 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9mfh\" (UniqueName: \"kubernetes.io/projected/b7aa3a14-1317-4a45-945b-4bb6eb26c9ea-kube-api-access-m9mfh\") pod \"route-controller-manager-94f88b9db-vth62\" (UID: \"b7aa3a14-1317-4a45-945b-4bb6eb26c9ea\") " pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.220591 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-67f9b8c6bd-smkkz" podStartSLOduration=7.220568254 podStartE2EDuration="7.220568254s" podCreationTimestamp="2025-12-06 14:02:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:02:16.215013915 +0000 UTC m=+379.050806802" watchObservedRunningTime="2025-12-06 14:02:16.220568254 +0000 UTC m=+379.056361151" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.348876 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" Dec 06 14:02:16 crc kubenswrapper[4706]: I1206 14:02:16.841357 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62"] Dec 06 14:02:17 crc kubenswrapper[4706]: I1206 14:02:17.205943 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" event={"ID":"b7aa3a14-1317-4a45-945b-4bb6eb26c9ea","Type":"ContainerStarted","Data":"baead59502a88a4880ada2fb277031466b81a9fd296c199f0233ddab2f3a93c5"} Dec 06 14:02:18 crc kubenswrapper[4706]: I1206 14:02:18.214602 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" event={"ID":"b7aa3a14-1317-4a45-945b-4bb6eb26c9ea","Type":"ContainerStarted","Data":"7eb11f17c9740221d1e6daa62e9d5faf43dd93cfea516045f7f433d4054f8baf"} Dec 06 14:02:18 crc kubenswrapper[4706]: I1206 14:02:18.235070 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" podStartSLOduration=8.235040502 podStartE2EDuration="8.235040502s" podCreationTimestamp="2025-12-06 14:02:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:02:18.234838127 +0000 UTC m=+381.070631024" watchObservedRunningTime="2025-12-06 14:02:18.235040502 +0000 UTC m=+381.070833389" Dec 06 14:02:19 crc kubenswrapper[4706]: I1206 14:02:19.220899 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" Dec 06 14:02:19 crc kubenswrapper[4706]: I1206 14:02:19.229573 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-94f88b9db-vth62" Dec 06 14:02:44 crc kubenswrapper[4706]: I1206 14:02:44.580214 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:02:44 crc kubenswrapper[4706]: I1206 14:02:44.580715 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.677184 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-kdlg7"] Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.678962 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.708561 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-kdlg7"] Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.862835 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/364d260e-b745-4a4a-98ff-8c11051156c1-bound-sa-token\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.863379 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.863525 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/364d260e-b745-4a4a-98ff-8c11051156c1-registry-tls\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.863718 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/364d260e-b745-4a4a-98ff-8c11051156c1-trusted-ca\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.863767 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/364d260e-b745-4a4a-98ff-8c11051156c1-registry-certificates\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.863994 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/364d260e-b745-4a4a-98ff-8c11051156c1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.864194 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/364d260e-b745-4a4a-98ff-8c11051156c1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.864246 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mgbd\" (UniqueName: \"kubernetes.io/projected/364d260e-b745-4a4a-98ff-8c11051156c1-kube-api-access-2mgbd\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.889672 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.965716 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/364d260e-b745-4a4a-98ff-8c11051156c1-trusted-ca\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.965772 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/364d260e-b745-4a4a-98ff-8c11051156c1-registry-certificates\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.965801 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/364d260e-b745-4a4a-98ff-8c11051156c1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.965843 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/364d260e-b745-4a4a-98ff-8c11051156c1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.965871 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mgbd\" (UniqueName: \"kubernetes.io/projected/364d260e-b745-4a4a-98ff-8c11051156c1-kube-api-access-2mgbd\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.965907 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/364d260e-b745-4a4a-98ff-8c11051156c1-bound-sa-token\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.965929 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/364d260e-b745-4a4a-98ff-8c11051156c1-registry-tls\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.967460 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/364d260e-b745-4a4a-98ff-8c11051156c1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.970769 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/364d260e-b745-4a4a-98ff-8c11051156c1-trusted-ca\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.971752 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/364d260e-b745-4a4a-98ff-8c11051156c1-registry-certificates\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.974662 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/364d260e-b745-4a4a-98ff-8c11051156c1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.977402 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/364d260e-b745-4a4a-98ff-8c11051156c1-registry-tls\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.986103 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mgbd\" (UniqueName: \"kubernetes.io/projected/364d260e-b745-4a4a-98ff-8c11051156c1-kube-api-access-2mgbd\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.989356 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/364d260e-b745-4a4a-98ff-8c11051156c1-bound-sa-token\") pod \"image-registry-66df7c8f76-kdlg7\" (UID: \"364d260e-b745-4a4a-98ff-8c11051156c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:00 crc kubenswrapper[4706]: I1206 14:03:00.999050 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:01 crc kubenswrapper[4706]: I1206 14:03:01.410880 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-kdlg7"] Dec 06 14:03:01 crc kubenswrapper[4706]: W1206 14:03:01.414767 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod364d260e_b745_4a4a_98ff_8c11051156c1.slice/crio-6d60bcfeeea65accc6b7cfa3e00094477ffb6e05ef52be02286dbee826fde354 WatchSource:0}: Error finding container 6d60bcfeeea65accc6b7cfa3e00094477ffb6e05ef52be02286dbee826fde354: Status 404 returned error can't find the container with id 6d60bcfeeea65accc6b7cfa3e00094477ffb6e05ef52be02286dbee826fde354 Dec 06 14:03:01 crc kubenswrapper[4706]: I1206 14:03:01.526241 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" event={"ID":"364d260e-b745-4a4a-98ff-8c11051156c1","Type":"ContainerStarted","Data":"6d60bcfeeea65accc6b7cfa3e00094477ffb6e05ef52be02286dbee826fde354"} Dec 06 14:03:02 crc kubenswrapper[4706]: I1206 14:03:02.534280 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" event={"ID":"364d260e-b745-4a4a-98ff-8c11051156c1","Type":"ContainerStarted","Data":"aa44f8eced666ddc440a096720f5ff4ca2b114f507714b1098fbca296ee66b0c"} Dec 06 14:03:02 crc kubenswrapper[4706]: I1206 14:03:02.534595 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:02 crc kubenswrapper[4706]: I1206 14:03:02.555613 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" podStartSLOduration=2.555594967 podStartE2EDuration="2.555594967s" podCreationTimestamp="2025-12-06 14:03:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:03:02.553433331 +0000 UTC m=+425.389226188" watchObservedRunningTime="2025-12-06 14:03:02.555594967 +0000 UTC m=+425.391387824" Dec 06 14:03:14 crc kubenswrapper[4706]: I1206 14:03:14.580771 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:03:14 crc kubenswrapper[4706]: I1206 14:03:14.581412 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:03:14 crc kubenswrapper[4706]: I1206 14:03:14.581480 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 14:03:14 crc kubenswrapper[4706]: I1206 14:03:14.582465 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"067a72c2e70239f7868cd95da2f6bc3f2b176f2e63176fd1a3892a751afb382e"} pod="openshift-machine-config-operator/machine-config-daemon-67v56" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 14:03:14 crc kubenswrapper[4706]: I1206 14:03:14.582577 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" containerID="cri-o://067a72c2e70239f7868cd95da2f6bc3f2b176f2e63176fd1a3892a751afb382e" gracePeriod=600 Dec 06 14:03:15 crc kubenswrapper[4706]: I1206 14:03:15.633731 4706 generic.go:334] "Generic (PLEG): container finished" podID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerID="067a72c2e70239f7868cd95da2f6bc3f2b176f2e63176fd1a3892a751afb382e" exitCode=0 Dec 06 14:03:15 crc kubenswrapper[4706]: I1206 14:03:15.633819 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerDied","Data":"067a72c2e70239f7868cd95da2f6bc3f2b176f2e63176fd1a3892a751afb382e"} Dec 06 14:03:15 crc kubenswrapper[4706]: I1206 14:03:15.634208 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerStarted","Data":"22caa6fab03e2ecbf77dfdaba33f02fdf7629d6e7b5d96a389d97d1f79320ebe"} Dec 06 14:03:15 crc kubenswrapper[4706]: I1206 14:03:15.634249 4706 scope.go:117] "RemoveContainer" containerID="244eeede55516bf0e9ada9fd040024472dce2dc077fd631baa2ec06ff80be41f" Dec 06 14:03:21 crc kubenswrapper[4706]: I1206 14:03:21.004262 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-kdlg7" Dec 06 14:03:21 crc kubenswrapper[4706]: I1206 14:03:21.066167 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-z4bwl"] Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.118101 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" podUID="ed999748-8e6d-4402-81e3-1f483b8440e8" containerName="registry" containerID="cri-o://164ee49b83f94279c8df627e642fa48664a0008fed4c3c0d55d6a6a729a656f7" gracePeriod=30 Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.570959 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.655188 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-registry-tls\") pod \"ed999748-8e6d-4402-81e3-1f483b8440e8\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.655256 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ed999748-8e6d-4402-81e3-1f483b8440e8-ca-trust-extracted\") pod \"ed999748-8e6d-4402-81e3-1f483b8440e8\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.662440 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "ed999748-8e6d-4402-81e3-1f483b8440e8" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.680721 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed999748-8e6d-4402-81e3-1f483b8440e8-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "ed999748-8e6d-4402-81e3-1f483b8440e8" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.756134 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ed999748-8e6d-4402-81e3-1f483b8440e8-installation-pull-secrets\") pod \"ed999748-8e6d-4402-81e3-1f483b8440e8\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.756414 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"ed999748-8e6d-4402-81e3-1f483b8440e8\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.756450 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l25h5\" (UniqueName: \"kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-kube-api-access-l25h5\") pod \"ed999748-8e6d-4402-81e3-1f483b8440e8\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.756494 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-bound-sa-token\") pod \"ed999748-8e6d-4402-81e3-1f483b8440e8\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.756521 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ed999748-8e6d-4402-81e3-1f483b8440e8-registry-certificates\") pod \"ed999748-8e6d-4402-81e3-1f483b8440e8\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.756648 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed999748-8e6d-4402-81e3-1f483b8440e8-trusted-ca\") pod \"ed999748-8e6d-4402-81e3-1f483b8440e8\" (UID: \"ed999748-8e6d-4402-81e3-1f483b8440e8\") " Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.756923 4706 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.756945 4706 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ed999748-8e6d-4402-81e3-1f483b8440e8-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.757257 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed999748-8e6d-4402-81e3-1f483b8440e8-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "ed999748-8e6d-4402-81e3-1f483b8440e8" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.757554 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed999748-8e6d-4402-81e3-1f483b8440e8-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "ed999748-8e6d-4402-81e3-1f483b8440e8" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.761450 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-kube-api-access-l25h5" (OuterVolumeSpecName: "kube-api-access-l25h5") pod "ed999748-8e6d-4402-81e3-1f483b8440e8" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8"). InnerVolumeSpecName "kube-api-access-l25h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.761882 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "ed999748-8e6d-4402-81e3-1f483b8440e8" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.764639 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed999748-8e6d-4402-81e3-1f483b8440e8-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "ed999748-8e6d-4402-81e3-1f483b8440e8" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.769603 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "ed999748-8e6d-4402-81e3-1f483b8440e8" (UID: "ed999748-8e6d-4402-81e3-1f483b8440e8"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.849750 4706 generic.go:334] "Generic (PLEG): container finished" podID="ed999748-8e6d-4402-81e3-1f483b8440e8" containerID="164ee49b83f94279c8df627e642fa48664a0008fed4c3c0d55d6a6a729a656f7" exitCode=0 Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.849808 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" event={"ID":"ed999748-8e6d-4402-81e3-1f483b8440e8","Type":"ContainerDied","Data":"164ee49b83f94279c8df627e642fa48664a0008fed4c3c0d55d6a6a729a656f7"} Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.849829 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.849869 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-z4bwl" event={"ID":"ed999748-8e6d-4402-81e3-1f483b8440e8","Type":"ContainerDied","Data":"de0693966858e44e25daaeb7d8a9be5818024e1c46f58dabd14c84a6ee4e9fe5"} Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.849894 4706 scope.go:117] "RemoveContainer" containerID="164ee49b83f94279c8df627e642fa48664a0008fed4c3c0d55d6a6a729a656f7" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.857975 4706 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ed999748-8e6d-4402-81e3-1f483b8440e8-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.858007 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l25h5\" (UniqueName: \"kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-kube-api-access-l25h5\") on node \"crc\" DevicePath \"\"" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.858021 4706 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ed999748-8e6d-4402-81e3-1f483b8440e8-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.858034 4706 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ed999748-8e6d-4402-81e3-1f483b8440e8-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.858046 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed999748-8e6d-4402-81e3-1f483b8440e8-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.875548 4706 scope.go:117] "RemoveContainer" containerID="164ee49b83f94279c8df627e642fa48664a0008fed4c3c0d55d6a6a729a656f7" Dec 06 14:03:46 crc kubenswrapper[4706]: E1206 14:03:46.876001 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"164ee49b83f94279c8df627e642fa48664a0008fed4c3c0d55d6a6a729a656f7\": container with ID starting with 164ee49b83f94279c8df627e642fa48664a0008fed4c3c0d55d6a6a729a656f7 not found: ID does not exist" containerID="164ee49b83f94279c8df627e642fa48664a0008fed4c3c0d55d6a6a729a656f7" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.876032 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"164ee49b83f94279c8df627e642fa48664a0008fed4c3c0d55d6a6a729a656f7"} err="failed to get container status \"164ee49b83f94279c8df627e642fa48664a0008fed4c3c0d55d6a6a729a656f7\": rpc error: code = NotFound desc = could not find container \"164ee49b83f94279c8df627e642fa48664a0008fed4c3c0d55d6a6a729a656f7\": container with ID starting with 164ee49b83f94279c8df627e642fa48664a0008fed4c3c0d55d6a6a729a656f7 not found: ID does not exist" Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.907148 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-z4bwl"] Dec 06 14:03:46 crc kubenswrapper[4706]: I1206 14:03:46.910762 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-z4bwl"] Dec 06 14:03:47 crc kubenswrapper[4706]: I1206 14:03:47.525245 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed999748-8e6d-4402-81e3-1f483b8440e8" path="/var/lib/kubelet/pods/ed999748-8e6d-4402-81e3-1f483b8440e8/volumes" Dec 06 14:05:14 crc kubenswrapper[4706]: I1206 14:05:14.580176 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:05:14 crc kubenswrapper[4706]: I1206 14:05:14.580923 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:05:44 crc kubenswrapper[4706]: I1206 14:05:44.592054 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:05:44 crc kubenswrapper[4706]: I1206 14:05:44.592930 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:06:14 crc kubenswrapper[4706]: I1206 14:06:14.580282 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:06:14 crc kubenswrapper[4706]: I1206 14:06:14.581276 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:06:14 crc kubenswrapper[4706]: I1206 14:06:14.581396 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 14:06:14 crc kubenswrapper[4706]: I1206 14:06:14.582139 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"22caa6fab03e2ecbf77dfdaba33f02fdf7629d6e7b5d96a389d97d1f79320ebe"} pod="openshift-machine-config-operator/machine-config-daemon-67v56" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 14:06:14 crc kubenswrapper[4706]: I1206 14:06:14.582237 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" containerID="cri-o://22caa6fab03e2ecbf77dfdaba33f02fdf7629d6e7b5d96a389d97d1f79320ebe" gracePeriod=600 Dec 06 14:06:14 crc kubenswrapper[4706]: I1206 14:06:14.851143 4706 generic.go:334] "Generic (PLEG): container finished" podID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerID="22caa6fab03e2ecbf77dfdaba33f02fdf7629d6e7b5d96a389d97d1f79320ebe" exitCode=0 Dec 06 14:06:14 crc kubenswrapper[4706]: I1206 14:06:14.851212 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerDied","Data":"22caa6fab03e2ecbf77dfdaba33f02fdf7629d6e7b5d96a389d97d1f79320ebe"} Dec 06 14:06:14 crc kubenswrapper[4706]: I1206 14:06:14.851271 4706 scope.go:117] "RemoveContainer" containerID="067a72c2e70239f7868cd95da2f6bc3f2b176f2e63176fd1a3892a751afb382e" Dec 06 14:06:15 crc kubenswrapper[4706]: I1206 14:06:15.861127 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerStarted","Data":"8f8d8c75a654551bee4406e15ba9924159f4ad79dd336b58107e1b41c26cbce8"} Dec 06 14:07:17 crc kubenswrapper[4706]: I1206 14:07:17.256985 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-q6bgc"] Dec 06 14:07:17 crc kubenswrapper[4706]: I1206 14:07:17.258590 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovn-controller" containerID="cri-o://ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09" gracePeriod=30 Dec 06 14:07:17 crc kubenswrapper[4706]: I1206 14:07:17.258703 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="northd" containerID="cri-o://d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f" gracePeriod=30 Dec 06 14:07:17 crc kubenswrapper[4706]: I1206 14:07:17.258673 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="nbdb" containerID="cri-o://af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a" gracePeriod=30 Dec 06 14:07:17 crc kubenswrapper[4706]: I1206 14:07:17.258697 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc" gracePeriod=30 Dec 06 14:07:17 crc kubenswrapper[4706]: I1206 14:07:17.259067 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="sbdb" containerID="cri-o://e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49" gracePeriod=30 Dec 06 14:07:17 crc kubenswrapper[4706]: I1206 14:07:17.258938 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="kube-rbac-proxy-node" containerID="cri-o://2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389" gracePeriod=30 Dec 06 14:07:17 crc kubenswrapper[4706]: I1206 14:07:17.263978 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovn-acl-logging" containerID="cri-o://6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9" gracePeriod=30 Dec 06 14:07:17 crc kubenswrapper[4706]: I1206 14:07:17.335508 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovnkube-controller" containerID="cri-o://0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7" gracePeriod=30 Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.035136 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.035222 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.037444 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.037502 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.039493 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.039614 4706 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="nbdb" Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.039657 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.039720 4706 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="sbdb" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.860816 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovnkube-controller/3.log" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.864884 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovn-acl-logging/0.log" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.866138 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovn-controller/0.log" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.867008 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.922592 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8nbns"] Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.926076 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovnkube-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926115 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovnkube-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.926124 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovnkube-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926129 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovnkube-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.926138 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="kubecfg-setup" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926144 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="kubecfg-setup" Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.926174 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovnkube-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926181 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovnkube-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.926190 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovn-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926197 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovn-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.926204 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovnkube-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926209 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovnkube-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.926216 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="kube-rbac-proxy-node" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926221 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="kube-rbac-proxy-node" Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.926229 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="nbdb" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926235 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="nbdb" Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.926259 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed999748-8e6d-4402-81e3-1f483b8440e8" containerName="registry" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926265 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed999748-8e6d-4402-81e3-1f483b8440e8" containerName="registry" Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.926274 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="northd" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926300 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="northd" Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.926331 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926337 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.926346 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovn-acl-logging" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926353 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovn-acl-logging" Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.926362 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="sbdb" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926368 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="sbdb" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926578 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="nbdb" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926680 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovnkube-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926691 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovnkube-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926698 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed999748-8e6d-4402-81e3-1f483b8440e8" containerName="registry" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926704 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="northd" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926712 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovn-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926719 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="kube-rbac-proxy-node" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926725 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovnkube-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926750 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovnkube-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926759 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="sbdb" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926766 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926775 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovn-acl-logging" Dec 06 14:07:18 crc kubenswrapper[4706]: E1206 14:07:18.926967 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovnkube-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.926977 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovnkube-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.927062 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerName="ovnkube-controller" Dec 06 14:07:18 crc kubenswrapper[4706]: I1206 14:07:18.928487 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.053378 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-openvswitch\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.053527 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.053622 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-systemd\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.053806 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-slash\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.053879 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-run-netns\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.053951 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovnkube-script-lib\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054025 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054055 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-cni-netd\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054084 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-var-lib-openvswitch\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054119 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-cni-bin\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.053945 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-slash" (OuterVolumeSpecName: "host-slash") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054147 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.053984 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054157 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnr6h\" (UniqueName: \"kubernetes.io/projected/c6db0d03-fbc0-4b3b-b89d-282767c274be-kube-api-access-fnr6h\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054196 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054229 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054247 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovnkube-config\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054238 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054381 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-kubelet\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054411 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-node-log\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054430 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054443 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-node-log" (OuterVolumeSpecName: "node-log") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054480 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-run-ovn-kubernetes\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054508 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-ovn\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054540 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovn-node-metrics-cert\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054559 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054599 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054604 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-systemd-units\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054664 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054726 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.055189 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.055213 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.054689 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-etc-openvswitch\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.056277 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-log-socket\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.056351 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-log-socket" (OuterVolumeSpecName: "log-socket") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.056385 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-env-overrides\") pod \"c6db0d03-fbc0-4b3b-b89d-282767c274be\" (UID: \"c6db0d03-fbc0-4b3b-b89d-282767c274be\") " Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.056686 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-run-systemd\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.056750 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.056806 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-run-netns\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.056837 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndf5v\" (UniqueName: \"kubernetes.io/projected/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-kube-api-access-ndf5v\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.056870 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-ovn-node-metrics-cert\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.056946 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-node-log\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.056980 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-run-openvswitch\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.057022 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-var-lib-openvswitch\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.057098 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-run-ovn-kubernetes\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.057133 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-cni-netd\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.057165 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-log-socket\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.057267 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-slash\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.057443 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.057504 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-ovnkube-config\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.057550 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-systemd-units\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.057586 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-cni-bin\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.057623 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-run-ovn\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.057657 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-env-overrides\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.057720 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-etc-openvswitch\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.057768 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-ovnkube-script-lib\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.057854 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-kubelet\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.058048 4706 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.058083 4706 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-log-socket\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.058103 4706 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.058122 4706 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.058141 4706 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-slash\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.058159 4706 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.058185 4706 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.058210 4706 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.058228 4706 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.058246 4706 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.058264 4706 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.058284 4706 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.058303 4706 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.058346 4706 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-node-log\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.058364 4706 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.058381 4706 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.058402 4706 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.063133 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6db0d03-fbc0-4b3b-b89d-282767c274be-kube-api-access-fnr6h" (OuterVolumeSpecName: "kube-api-access-fnr6h") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "kube-api-access-fnr6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.063238 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.069412 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "c6db0d03-fbc0-4b3b-b89d-282767c274be" (UID: "c6db0d03-fbc0-4b3b-b89d-282767c274be"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.159838 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-ovn-node-metrics-cert\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.159906 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-node-log\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.159935 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-run-openvswitch\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.159967 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-var-lib-openvswitch\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160021 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-run-ovn-kubernetes\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160048 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-cni-netd\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160078 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-log-socket\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160079 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-node-log\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160101 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-slash\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160149 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-slash\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160162 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160189 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160201 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-ovnkube-config\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160216 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-run-ovn-kubernetes\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160231 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-systemd-units\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160244 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-run-openvswitch\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160258 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-cni-bin\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160266 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-var-lib-openvswitch\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160282 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-run-ovn\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160348 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-env-overrides\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160380 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-etc-openvswitch\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160409 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-ovnkube-script-lib\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160438 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-kubelet\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160480 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-run-systemd\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160545 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-run-netns\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160568 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndf5v\" (UniqueName: \"kubernetes.io/projected/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-kube-api-access-ndf5v\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160624 4706 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6db0d03-fbc0-4b3b-b89d-282767c274be-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160641 4706 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6db0d03-fbc0-4b3b-b89d-282767c274be-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160656 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnr6h\" (UniqueName: \"kubernetes.io/projected/c6db0d03-fbc0-4b3b-b89d-282767c274be-kube-api-access-fnr6h\") on node \"crc\" DevicePath \"\"" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160851 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-cni-netd\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160944 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-log-socket\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.161021 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-systemd-units\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.161023 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-etc-openvswitch\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.161062 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-cni-bin\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.161083 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-run-ovn\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.160989 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-ovnkube-config\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.161519 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-run-systemd\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.161561 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-env-overrides\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.161585 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-kubelet\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.161602 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-host-run-netns\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.161808 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-ovnkube-script-lib\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.164220 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-ovn-node-metrics-cert\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.178999 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndf5v\" (UniqueName: \"kubernetes.io/projected/fc83d95d-ebdd-4816-8a2e-878ef23a8aac-kube-api-access-ndf5v\") pod \"ovnkube-node-8nbns\" (UID: \"fc83d95d-ebdd-4816-8a2e-878ef23a8aac\") " pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.244702 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.294059 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" event={"ID":"fc83d95d-ebdd-4816-8a2e-878ef23a8aac","Type":"ContainerStarted","Data":"6c2eb28376b35ba5110deb9c67e951cf205cf2f2be3b3d51de40897058ae0e91"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.296879 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovnkube-controller/3.log" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.301069 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovn-acl-logging/0.log" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.301832 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q6bgc_c6db0d03-fbc0-4b3b-b89d-282767c274be/ovn-controller/0.log" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302207 4706 generic.go:334] "Generic (PLEG): container finished" podID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerID="0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7" exitCode=0 Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302238 4706 generic.go:334] "Generic (PLEG): container finished" podID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerID="e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49" exitCode=0 Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302249 4706 generic.go:334] "Generic (PLEG): container finished" podID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerID="af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a" exitCode=0 Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302259 4706 generic.go:334] "Generic (PLEG): container finished" podID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerID="d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f" exitCode=0 Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302270 4706 generic.go:334] "Generic (PLEG): container finished" podID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerID="42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc" exitCode=0 Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302279 4706 generic.go:334] "Generic (PLEG): container finished" podID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerID="2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389" exitCode=0 Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302287 4706 generic.go:334] "Generic (PLEG): container finished" podID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerID="6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9" exitCode=143 Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302297 4706 generic.go:334] "Generic (PLEG): container finished" podID="c6db0d03-fbc0-4b3b-b89d-282767c274be" containerID="ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09" exitCode=143 Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302374 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerDied","Data":"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302413 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerDied","Data":"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302431 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerDied","Data":"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302444 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerDied","Data":"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302457 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerDied","Data":"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302473 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerDied","Data":"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302488 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302502 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302510 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302517 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302523 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302530 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302542 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302549 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302555 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302564 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerDied","Data":"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302576 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302585 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302592 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302598 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302605 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302612 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302620 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302626 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302633 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302640 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302649 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerDied","Data":"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302663 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302672 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302679 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302686 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302693 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302700 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302707 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302714 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302722 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302728 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302739 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" event={"ID":"c6db0d03-fbc0-4b3b-b89d-282767c274be","Type":"ContainerDied","Data":"52e8d9434c7c56ec1bb8c5241162afeca03b248ddca4b5d72091792df4a71ec0"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302752 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302761 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302769 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302776 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302784 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302791 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302798 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302805 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302812 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302818 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.302838 4706 scope.go:117] "RemoveContainer" containerID="0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.303020 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-q6bgc" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.315020 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xqlx6_d3fdd6dc-f817-486c-b0ee-18a6f2185559/kube-multus/2.log" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.315719 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xqlx6_d3fdd6dc-f817-486c-b0ee-18a6f2185559/kube-multus/1.log" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.315754 4706 generic.go:334] "Generic (PLEG): container finished" podID="d3fdd6dc-f817-486c-b0ee-18a6f2185559" containerID="eaf8244ef2489d4736fe70838f4ddad49ff130d448cd95eac58babc9d24ad22b" exitCode=2 Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.315779 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xqlx6" event={"ID":"d3fdd6dc-f817-486c-b0ee-18a6f2185559","Type":"ContainerDied","Data":"eaf8244ef2489d4736fe70838f4ddad49ff130d448cd95eac58babc9d24ad22b"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.315797 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fcaf233bcb76d165e238dfbafde4bcbfa99d5a083980078c23ea17e2ab19d5af"} Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.316142 4706 scope.go:117] "RemoveContainer" containerID="eaf8244ef2489d4736fe70838f4ddad49ff130d448cd95eac58babc9d24ad22b" Dec 06 14:07:19 crc kubenswrapper[4706]: E1206 14:07:19.316308 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-xqlx6_openshift-multus(d3fdd6dc-f817-486c-b0ee-18a6f2185559)\"" pod="openshift-multus/multus-xqlx6" podUID="d3fdd6dc-f817-486c-b0ee-18a6f2185559" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.367404 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-q6bgc"] Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.368160 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-q6bgc"] Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.371519 4706 scope.go:117] "RemoveContainer" containerID="aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.388551 4706 scope.go:117] "RemoveContainer" containerID="e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.418725 4706 scope.go:117] "RemoveContainer" containerID="af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.471471 4706 scope.go:117] "RemoveContainer" containerID="d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.489404 4706 scope.go:117] "RemoveContainer" containerID="42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.506993 4706 scope.go:117] "RemoveContainer" containerID="2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.520061 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6db0d03-fbc0-4b3b-b89d-282767c274be" path="/var/lib/kubelet/pods/c6db0d03-fbc0-4b3b-b89d-282767c274be/volumes" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.528747 4706 scope.go:117] "RemoveContainer" containerID="6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.547185 4706 scope.go:117] "RemoveContainer" containerID="ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.567111 4706 scope.go:117] "RemoveContainer" containerID="8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.587855 4706 scope.go:117] "RemoveContainer" containerID="0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7" Dec 06 14:07:19 crc kubenswrapper[4706]: E1206 14:07:19.588514 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7\": container with ID starting with 0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7 not found: ID does not exist" containerID="0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.588545 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7"} err="failed to get container status \"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7\": rpc error: code = NotFound desc = could not find container \"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7\": container with ID starting with 0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.588569 4706 scope.go:117] "RemoveContainer" containerID="aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17" Dec 06 14:07:19 crc kubenswrapper[4706]: E1206 14:07:19.589014 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17\": container with ID starting with aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17 not found: ID does not exist" containerID="aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.589035 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17"} err="failed to get container status \"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17\": rpc error: code = NotFound desc = could not find container \"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17\": container with ID starting with aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.589047 4706 scope.go:117] "RemoveContainer" containerID="e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49" Dec 06 14:07:19 crc kubenswrapper[4706]: E1206 14:07:19.589486 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\": container with ID starting with e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49 not found: ID does not exist" containerID="e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.589508 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49"} err="failed to get container status \"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\": rpc error: code = NotFound desc = could not find container \"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\": container with ID starting with e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.589520 4706 scope.go:117] "RemoveContainer" containerID="af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a" Dec 06 14:07:19 crc kubenswrapper[4706]: E1206 14:07:19.589928 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\": container with ID starting with af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a not found: ID does not exist" containerID="af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.589983 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a"} err="failed to get container status \"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\": rpc error: code = NotFound desc = could not find container \"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\": container with ID starting with af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.590015 4706 scope.go:117] "RemoveContainer" containerID="d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f" Dec 06 14:07:19 crc kubenswrapper[4706]: E1206 14:07:19.590601 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\": container with ID starting with d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f not found: ID does not exist" containerID="d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.590629 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f"} err="failed to get container status \"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\": rpc error: code = NotFound desc = could not find container \"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\": container with ID starting with d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.590642 4706 scope.go:117] "RemoveContainer" containerID="42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc" Dec 06 14:07:19 crc kubenswrapper[4706]: E1206 14:07:19.590940 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\": container with ID starting with 42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc not found: ID does not exist" containerID="42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.590963 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc"} err="failed to get container status \"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\": rpc error: code = NotFound desc = could not find container \"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\": container with ID starting with 42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.590975 4706 scope.go:117] "RemoveContainer" containerID="2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389" Dec 06 14:07:19 crc kubenswrapper[4706]: E1206 14:07:19.591215 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\": container with ID starting with 2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389 not found: ID does not exist" containerID="2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.591234 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389"} err="failed to get container status \"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\": rpc error: code = NotFound desc = could not find container \"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\": container with ID starting with 2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.591247 4706 scope.go:117] "RemoveContainer" containerID="6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9" Dec 06 14:07:19 crc kubenswrapper[4706]: E1206 14:07:19.591830 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\": container with ID starting with 6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9 not found: ID does not exist" containerID="6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.592754 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9"} err="failed to get container status \"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\": rpc error: code = NotFound desc = could not find container \"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\": container with ID starting with 6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.592818 4706 scope.go:117] "RemoveContainer" containerID="ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09" Dec 06 14:07:19 crc kubenswrapper[4706]: E1206 14:07:19.593253 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\": container with ID starting with ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09 not found: ID does not exist" containerID="ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.593280 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09"} err="failed to get container status \"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\": rpc error: code = NotFound desc = could not find container \"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\": container with ID starting with ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.593292 4706 scope.go:117] "RemoveContainer" containerID="8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773" Dec 06 14:07:19 crc kubenswrapper[4706]: E1206 14:07:19.593611 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\": container with ID starting with 8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773 not found: ID does not exist" containerID="8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.593628 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773"} err="failed to get container status \"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\": rpc error: code = NotFound desc = could not find container \"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\": container with ID starting with 8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.593646 4706 scope.go:117] "RemoveContainer" containerID="0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.593901 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7"} err="failed to get container status \"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7\": rpc error: code = NotFound desc = could not find container \"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7\": container with ID starting with 0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.593931 4706 scope.go:117] "RemoveContainer" containerID="aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.594287 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17"} err="failed to get container status \"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17\": rpc error: code = NotFound desc = could not find container \"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17\": container with ID starting with aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.594341 4706 scope.go:117] "RemoveContainer" containerID="e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.594592 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49"} err="failed to get container status \"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\": rpc error: code = NotFound desc = could not find container \"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\": container with ID starting with e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.594610 4706 scope.go:117] "RemoveContainer" containerID="af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.594825 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a"} err="failed to get container status \"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\": rpc error: code = NotFound desc = could not find container \"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\": container with ID starting with af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.594848 4706 scope.go:117] "RemoveContainer" containerID="d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.595339 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f"} err="failed to get container status \"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\": rpc error: code = NotFound desc = could not find container \"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\": container with ID starting with d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.595359 4706 scope.go:117] "RemoveContainer" containerID="42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.596138 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc"} err="failed to get container status \"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\": rpc error: code = NotFound desc = could not find container \"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\": container with ID starting with 42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.596168 4706 scope.go:117] "RemoveContainer" containerID="2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.596477 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389"} err="failed to get container status \"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\": rpc error: code = NotFound desc = could not find container \"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\": container with ID starting with 2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.596498 4706 scope.go:117] "RemoveContainer" containerID="6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.596750 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9"} err="failed to get container status \"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\": rpc error: code = NotFound desc = could not find container \"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\": container with ID starting with 6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.596773 4706 scope.go:117] "RemoveContainer" containerID="ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.597342 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09"} err="failed to get container status \"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\": rpc error: code = NotFound desc = could not find container \"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\": container with ID starting with ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.597373 4706 scope.go:117] "RemoveContainer" containerID="8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.597808 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773"} err="failed to get container status \"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\": rpc error: code = NotFound desc = could not find container \"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\": container with ID starting with 8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.597829 4706 scope.go:117] "RemoveContainer" containerID="0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.598273 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7"} err="failed to get container status \"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7\": rpc error: code = NotFound desc = could not find container \"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7\": container with ID starting with 0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.598300 4706 scope.go:117] "RemoveContainer" containerID="aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.598718 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17"} err="failed to get container status \"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17\": rpc error: code = NotFound desc = could not find container \"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17\": container with ID starting with aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.598741 4706 scope.go:117] "RemoveContainer" containerID="e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.599141 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49"} err="failed to get container status \"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\": rpc error: code = NotFound desc = could not find container \"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\": container with ID starting with e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.599162 4706 scope.go:117] "RemoveContainer" containerID="af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.599615 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a"} err="failed to get container status \"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\": rpc error: code = NotFound desc = could not find container \"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\": container with ID starting with af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.599647 4706 scope.go:117] "RemoveContainer" containerID="d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.599899 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f"} err="failed to get container status \"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\": rpc error: code = NotFound desc = could not find container \"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\": container with ID starting with d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.599926 4706 scope.go:117] "RemoveContainer" containerID="42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.600218 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc"} err="failed to get container status \"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\": rpc error: code = NotFound desc = could not find container \"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\": container with ID starting with 42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.600236 4706 scope.go:117] "RemoveContainer" containerID="2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.600503 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389"} err="failed to get container status \"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\": rpc error: code = NotFound desc = could not find container \"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\": container with ID starting with 2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.600525 4706 scope.go:117] "RemoveContainer" containerID="6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.600741 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9"} err="failed to get container status \"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\": rpc error: code = NotFound desc = could not find container \"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\": container with ID starting with 6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.600761 4706 scope.go:117] "RemoveContainer" containerID="ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.601367 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09"} err="failed to get container status \"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\": rpc error: code = NotFound desc = could not find container \"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\": container with ID starting with ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.601423 4706 scope.go:117] "RemoveContainer" containerID="8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.601880 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773"} err="failed to get container status \"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\": rpc error: code = NotFound desc = could not find container \"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\": container with ID starting with 8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.601903 4706 scope.go:117] "RemoveContainer" containerID="0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.602135 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7"} err="failed to get container status \"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7\": rpc error: code = NotFound desc = could not find container \"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7\": container with ID starting with 0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.602154 4706 scope.go:117] "RemoveContainer" containerID="aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.602389 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17"} err="failed to get container status \"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17\": rpc error: code = NotFound desc = could not find container \"aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17\": container with ID starting with aee67015063dbc43c985ddb04ee4ee12b230e11bffb83f7706d5b081ab00bd17 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.602414 4706 scope.go:117] "RemoveContainer" containerID="e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.602612 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49"} err="failed to get container status \"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\": rpc error: code = NotFound desc = could not find container \"e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49\": container with ID starting with e51b44d4ba92a6abc292276de7f0a6f100f3b5291be7533dd3823db85350cf49 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.602631 4706 scope.go:117] "RemoveContainer" containerID="af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.602805 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a"} err="failed to get container status \"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\": rpc error: code = NotFound desc = could not find container \"af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a\": container with ID starting with af92b2664ceb799dbe91ed122736c6ef9ad1c4144f222e6370e61b772a1b7d1a not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.602822 4706 scope.go:117] "RemoveContainer" containerID="d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.603131 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f"} err="failed to get container status \"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\": rpc error: code = NotFound desc = could not find container \"d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f\": container with ID starting with d73cb14b4c8beef5d8516a6d041c933636b7d554724f232f16d4548ed656366f not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.603152 4706 scope.go:117] "RemoveContainer" containerID="42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.603447 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc"} err="failed to get container status \"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\": rpc error: code = NotFound desc = could not find container \"42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc\": container with ID starting with 42a47acaab3fed86154a344201cfae75d0166887edce8e78baa12631d4a4c0dc not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.603464 4706 scope.go:117] "RemoveContainer" containerID="2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.603684 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389"} err="failed to get container status \"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\": rpc error: code = NotFound desc = could not find container \"2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389\": container with ID starting with 2713aafb6cdfcd635a451374d99ac4776345579fe1f9b09d05adfa70d1a3f389 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.603713 4706 scope.go:117] "RemoveContainer" containerID="6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.603952 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9"} err="failed to get container status \"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\": rpc error: code = NotFound desc = could not find container \"6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9\": container with ID starting with 6a178f705cb5ef148107c659e20f426338cc2704f8b1770fab63af9906637ae9 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.603971 4706 scope.go:117] "RemoveContainer" containerID="ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.604468 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09"} err="failed to get container status \"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\": rpc error: code = NotFound desc = could not find container \"ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09\": container with ID starting with ecf711a9d74ede5582b554dcc160d19fa4913e037a4dc4df32fbd49c01851f09 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.604487 4706 scope.go:117] "RemoveContainer" containerID="8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.604708 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773"} err="failed to get container status \"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\": rpc error: code = NotFound desc = could not find container \"8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773\": container with ID starting with 8b79383b667e684253930dd99a9791dd0c355e7c3edfacf11151c0e9273b3773 not found: ID does not exist" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.604733 4706 scope.go:117] "RemoveContainer" containerID="0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7" Dec 06 14:07:19 crc kubenswrapper[4706]: I1206 14:07:19.604939 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7"} err="failed to get container status \"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7\": rpc error: code = NotFound desc = could not find container \"0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7\": container with ID starting with 0635e5e924cc99f03e1f20d8f84cf97a9093e42a6f0a76ad4b9ba26502ded8e7 not found: ID does not exist" Dec 06 14:07:20 crc kubenswrapper[4706]: I1206 14:07:20.327944 4706 generic.go:334] "Generic (PLEG): container finished" podID="fc83d95d-ebdd-4816-8a2e-878ef23a8aac" containerID="bbf4df332b59fce2f963a819d8bce2672de82f2bb8c95cc3d1a92c5c05a04019" exitCode=0 Dec 06 14:07:20 crc kubenswrapper[4706]: I1206 14:07:20.328025 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" event={"ID":"fc83d95d-ebdd-4816-8a2e-878ef23a8aac","Type":"ContainerDied","Data":"bbf4df332b59fce2f963a819d8bce2672de82f2bb8c95cc3d1a92c5c05a04019"} Dec 06 14:07:21 crc kubenswrapper[4706]: I1206 14:07:21.339054 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" event={"ID":"fc83d95d-ebdd-4816-8a2e-878ef23a8aac","Type":"ContainerStarted","Data":"511e32210c77350cd8d3ca28bed894c388b4aa7b8b1119cd997be4eea13cf41d"} Dec 06 14:07:21 crc kubenswrapper[4706]: I1206 14:07:21.339471 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" event={"ID":"fc83d95d-ebdd-4816-8a2e-878ef23a8aac","Type":"ContainerStarted","Data":"1e01a8984211685fe43b78fa5dd015de86781e8ec89b647951ba6cee6f1f8253"} Dec 06 14:07:21 crc kubenswrapper[4706]: I1206 14:07:21.339493 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" event={"ID":"fc83d95d-ebdd-4816-8a2e-878ef23a8aac","Type":"ContainerStarted","Data":"9f718381cdd6f3532bf2c29db0c0d91c2067418e43012c04a66559652565f61e"} Dec 06 14:07:21 crc kubenswrapper[4706]: I1206 14:07:21.339512 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" event={"ID":"fc83d95d-ebdd-4816-8a2e-878ef23a8aac","Type":"ContainerStarted","Data":"ac88eef235b3d43a0b91af520247c2657bb55c3e6527da5ffd8c4c402a3af45f"} Dec 06 14:07:21 crc kubenswrapper[4706]: I1206 14:07:21.339530 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" event={"ID":"fc83d95d-ebdd-4816-8a2e-878ef23a8aac","Type":"ContainerStarted","Data":"9a0b0eddcd044a724116c27c75a7736b2d35a6665f830d731433384e9b8281ae"} Dec 06 14:07:21 crc kubenswrapper[4706]: I1206 14:07:21.339548 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" event={"ID":"fc83d95d-ebdd-4816-8a2e-878ef23a8aac","Type":"ContainerStarted","Data":"ab42032ced50b7fd646745b0f29885f28627b05b46678326862e4b3d44f4991c"} Dec 06 14:07:24 crc kubenswrapper[4706]: I1206 14:07:24.367205 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" event={"ID":"fc83d95d-ebdd-4816-8a2e-878ef23a8aac","Type":"ContainerStarted","Data":"8f167639d310af952dd7818861ba6d0298baf82913c679d663a39aa7f21bb98f"} Dec 06 14:07:26 crc kubenswrapper[4706]: I1206 14:07:26.385067 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" event={"ID":"fc83d95d-ebdd-4816-8a2e-878ef23a8aac","Type":"ContainerStarted","Data":"af9c27327bbc18c5fcbb47812bbd0e8467dd1bec3d7b4a38da243d6d820cf143"} Dec 06 14:07:26 crc kubenswrapper[4706]: I1206 14:07:26.385630 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:26 crc kubenswrapper[4706]: I1206 14:07:26.385647 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:26 crc kubenswrapper[4706]: I1206 14:07:26.385659 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:26 crc kubenswrapper[4706]: I1206 14:07:26.420480 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" podStartSLOduration=8.420457398 podStartE2EDuration="8.420457398s" podCreationTimestamp="2025-12-06 14:07:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:07:26.418149297 +0000 UTC m=+689.253942144" watchObservedRunningTime="2025-12-06 14:07:26.420457398 +0000 UTC m=+689.256250255" Dec 06 14:07:26 crc kubenswrapper[4706]: I1206 14:07:26.422743 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:26 crc kubenswrapper[4706]: I1206 14:07:26.432001 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:33 crc kubenswrapper[4706]: I1206 14:07:33.511132 4706 scope.go:117] "RemoveContainer" containerID="eaf8244ef2489d4736fe70838f4ddad49ff130d448cd95eac58babc9d24ad22b" Dec 06 14:07:33 crc kubenswrapper[4706]: E1206 14:07:33.512031 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-xqlx6_openshift-multus(d3fdd6dc-f817-486c-b0ee-18a6f2185559)\"" pod="openshift-multus/multus-xqlx6" podUID="d3fdd6dc-f817-486c-b0ee-18a6f2185559" Dec 06 14:07:43 crc kubenswrapper[4706]: I1206 14:07:43.562526 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s"] Dec 06 14:07:43 crc kubenswrapper[4706]: I1206 14:07:43.564292 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:43 crc kubenswrapper[4706]: I1206 14:07:43.567104 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 14:07:43 crc kubenswrapper[4706]: I1206 14:07:43.570186 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s"] Dec 06 14:07:43 crc kubenswrapper[4706]: I1206 14:07:43.713887 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s\" (UID: \"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:43 crc kubenswrapper[4706]: I1206 14:07:43.714034 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s\" (UID: \"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:43 crc kubenswrapper[4706]: I1206 14:07:43.714374 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dq6rn\" (UniqueName: \"kubernetes.io/projected/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-kube-api-access-dq6rn\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s\" (UID: \"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:43 crc kubenswrapper[4706]: I1206 14:07:43.818469 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dq6rn\" (UniqueName: \"kubernetes.io/projected/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-kube-api-access-dq6rn\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s\" (UID: \"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:43 crc kubenswrapper[4706]: I1206 14:07:43.818582 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s\" (UID: \"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:43 crc kubenswrapper[4706]: I1206 14:07:43.818653 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s\" (UID: \"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:43 crc kubenswrapper[4706]: I1206 14:07:43.819632 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s\" (UID: \"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:43 crc kubenswrapper[4706]: I1206 14:07:43.820162 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s\" (UID: \"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:43 crc kubenswrapper[4706]: I1206 14:07:43.847805 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dq6rn\" (UniqueName: \"kubernetes.io/projected/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-kube-api-access-dq6rn\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s\" (UID: \"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:43 crc kubenswrapper[4706]: I1206 14:07:43.887142 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:43 crc kubenswrapper[4706]: E1206 14:07:43.930246 4706 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_openshift-marketplace_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4_0(24655aa5efb5864974cf1818e0fe3a77562a8d0bb321b891ea86571e6ecde5f6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 14:07:43 crc kubenswrapper[4706]: E1206 14:07:43.930435 4706 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_openshift-marketplace_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4_0(24655aa5efb5864974cf1818e0fe3a77562a8d0bb321b891ea86571e6ecde5f6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:43 crc kubenswrapper[4706]: E1206 14:07:43.930487 4706 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_openshift-marketplace_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4_0(24655aa5efb5864974cf1818e0fe3a77562a8d0bb321b891ea86571e6ecde5f6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:43 crc kubenswrapper[4706]: E1206 14:07:43.930588 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_openshift-marketplace(51fa29f5-f7d7-4ac2-a15a-2b16025a00b4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_openshift-marketplace(51fa29f5-f7d7-4ac2-a15a-2b16025a00b4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_openshift-marketplace_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4_0(24655aa5efb5864974cf1818e0fe3a77562a8d0bb321b891ea86571e6ecde5f6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" podUID="51fa29f5-f7d7-4ac2-a15a-2b16025a00b4" Dec 06 14:07:44 crc kubenswrapper[4706]: I1206 14:07:44.514770 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:44 crc kubenswrapper[4706]: I1206 14:07:44.515565 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:44 crc kubenswrapper[4706]: E1206 14:07:44.553501 4706 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_openshift-marketplace_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4_0(bf4c1f62c90a1cdfdfcdfacc53d20937f214555b403cbea5b128c72323d327ab): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 14:07:44 crc kubenswrapper[4706]: E1206 14:07:44.553676 4706 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_openshift-marketplace_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4_0(bf4c1f62c90a1cdfdfcdfacc53d20937f214555b403cbea5b128c72323d327ab): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:44 crc kubenswrapper[4706]: E1206 14:07:44.553749 4706 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_openshift-marketplace_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4_0(bf4c1f62c90a1cdfdfcdfacc53d20937f214555b403cbea5b128c72323d327ab): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:44 crc kubenswrapper[4706]: E1206 14:07:44.553976 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_openshift-marketplace(51fa29f5-f7d7-4ac2-a15a-2b16025a00b4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_openshift-marketplace(51fa29f5-f7d7-4ac2-a15a-2b16025a00b4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_openshift-marketplace_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4_0(bf4c1f62c90a1cdfdfcdfacc53d20937f214555b403cbea5b128c72323d327ab): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" podUID="51fa29f5-f7d7-4ac2-a15a-2b16025a00b4" Dec 06 14:07:48 crc kubenswrapper[4706]: I1206 14:07:48.511467 4706 scope.go:117] "RemoveContainer" containerID="eaf8244ef2489d4736fe70838f4ddad49ff130d448cd95eac58babc9d24ad22b" Dec 06 14:07:49 crc kubenswrapper[4706]: I1206 14:07:49.276406 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8nbns" Dec 06 14:07:49 crc kubenswrapper[4706]: I1206 14:07:49.547180 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xqlx6_d3fdd6dc-f817-486c-b0ee-18a6f2185559/kube-multus/2.log" Dec 06 14:07:49 crc kubenswrapper[4706]: I1206 14:07:49.547707 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xqlx6_d3fdd6dc-f817-486c-b0ee-18a6f2185559/kube-multus/1.log" Dec 06 14:07:49 crc kubenswrapper[4706]: I1206 14:07:49.547771 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xqlx6" event={"ID":"d3fdd6dc-f817-486c-b0ee-18a6f2185559","Type":"ContainerStarted","Data":"b906b1ea815c47210c2bc21af6c14e71ffb9e81deccd8f975b2b221b63ffcf7f"} Dec 06 14:07:57 crc kubenswrapper[4706]: I1206 14:07:57.510413 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:57 crc kubenswrapper[4706]: I1206 14:07:57.516855 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:07:57 crc kubenswrapper[4706]: I1206 14:07:57.760452 4706 scope.go:117] "RemoveContainer" containerID="fcaf233bcb76d165e238dfbafde4bcbfa99d5a083980078c23ea17e2ab19d5af" Dec 06 14:07:57 crc kubenswrapper[4706]: I1206 14:07:57.800566 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s"] Dec 06 14:07:58 crc kubenswrapper[4706]: I1206 14:07:58.626912 4706 generic.go:334] "Generic (PLEG): container finished" podID="51fa29f5-f7d7-4ac2-a15a-2b16025a00b4" containerID="35842ebda16353aa664bdca30973c7f70e46ea4e18d421382b70fdb908a75ffe" exitCode=0 Dec 06 14:07:58 crc kubenswrapper[4706]: I1206 14:07:58.627006 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" event={"ID":"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4","Type":"ContainerDied","Data":"35842ebda16353aa664bdca30973c7f70e46ea4e18d421382b70fdb908a75ffe"} Dec 06 14:07:58 crc kubenswrapper[4706]: I1206 14:07:58.627043 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" event={"ID":"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4","Type":"ContainerStarted","Data":"2ff148b5f9e1f1a22eaa0ebc6d41334925332c240703f809420dd8ab35e24d77"} Dec 06 14:07:58 crc kubenswrapper[4706]: I1206 14:07:58.629465 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 14:07:58 crc kubenswrapper[4706]: I1206 14:07:58.631619 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xqlx6_d3fdd6dc-f817-486c-b0ee-18a6f2185559/kube-multus/2.log" Dec 06 14:08:00 crc kubenswrapper[4706]: I1206 14:08:00.649497 4706 generic.go:334] "Generic (PLEG): container finished" podID="51fa29f5-f7d7-4ac2-a15a-2b16025a00b4" containerID="db8162454fd2fe5879d4bde17bfa3ae583a6eef9b721940b2c633667daea7718" exitCode=0 Dec 06 14:08:00 crc kubenswrapper[4706]: I1206 14:08:00.649636 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" event={"ID":"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4","Type":"ContainerDied","Data":"db8162454fd2fe5879d4bde17bfa3ae583a6eef9b721940b2c633667daea7718"} Dec 06 14:08:01 crc kubenswrapper[4706]: I1206 14:08:01.658962 4706 generic.go:334] "Generic (PLEG): container finished" podID="51fa29f5-f7d7-4ac2-a15a-2b16025a00b4" containerID="537a5cb4f13051705a07321920da5aa050595929db8934ac27570676c445d16c" exitCode=0 Dec 06 14:08:01 crc kubenswrapper[4706]: I1206 14:08:01.659031 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" event={"ID":"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4","Type":"ContainerDied","Data":"537a5cb4f13051705a07321920da5aa050595929db8934ac27570676c445d16c"} Dec 06 14:08:02 crc kubenswrapper[4706]: I1206 14:08:02.973349 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:08:03 crc kubenswrapper[4706]: I1206 14:08:03.089374 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-util\") pod \"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4\" (UID: \"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4\") " Dec 06 14:08:03 crc kubenswrapper[4706]: I1206 14:08:03.089451 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-bundle\") pod \"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4\" (UID: \"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4\") " Dec 06 14:08:03 crc kubenswrapper[4706]: I1206 14:08:03.089559 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dq6rn\" (UniqueName: \"kubernetes.io/projected/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-kube-api-access-dq6rn\") pod \"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4\" (UID: \"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4\") " Dec 06 14:08:03 crc kubenswrapper[4706]: I1206 14:08:03.091256 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-bundle" (OuterVolumeSpecName: "bundle") pod "51fa29f5-f7d7-4ac2-a15a-2b16025a00b4" (UID: "51fa29f5-f7d7-4ac2-a15a-2b16025a00b4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:08:03 crc kubenswrapper[4706]: I1206 14:08:03.099611 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-kube-api-access-dq6rn" (OuterVolumeSpecName: "kube-api-access-dq6rn") pod "51fa29f5-f7d7-4ac2-a15a-2b16025a00b4" (UID: "51fa29f5-f7d7-4ac2-a15a-2b16025a00b4"). InnerVolumeSpecName "kube-api-access-dq6rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:08:03 crc kubenswrapper[4706]: I1206 14:08:03.107104 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-util" (OuterVolumeSpecName: "util") pod "51fa29f5-f7d7-4ac2-a15a-2b16025a00b4" (UID: "51fa29f5-f7d7-4ac2-a15a-2b16025a00b4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:08:03 crc kubenswrapper[4706]: I1206 14:08:03.191932 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dq6rn\" (UniqueName: \"kubernetes.io/projected/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-kube-api-access-dq6rn\") on node \"crc\" DevicePath \"\"" Dec 06 14:08:03 crc kubenswrapper[4706]: I1206 14:08:03.191987 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-util\") on node \"crc\" DevicePath \"\"" Dec 06 14:08:03 crc kubenswrapper[4706]: I1206 14:08:03.192007 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/51fa29f5-f7d7-4ac2-a15a-2b16025a00b4-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 14:08:03 crc kubenswrapper[4706]: I1206 14:08:03.678071 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" event={"ID":"51fa29f5-f7d7-4ac2-a15a-2b16025a00b4","Type":"ContainerDied","Data":"2ff148b5f9e1f1a22eaa0ebc6d41334925332c240703f809420dd8ab35e24d77"} Dec 06 14:08:03 crc kubenswrapper[4706]: I1206 14:08:03.678156 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ff148b5f9e1f1a22eaa0ebc6d41334925332c240703f809420dd8ab35e24d77" Dec 06 14:08:03 crc kubenswrapper[4706]: I1206 14:08:03.678381 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s" Dec 06 14:08:14 crc kubenswrapper[4706]: I1206 14:08:14.580471 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:08:14 crc kubenswrapper[4706]: I1206 14:08:14.581165 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.078074 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9"] Dec 06 14:08:17 crc kubenswrapper[4706]: E1206 14:08:17.078428 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51fa29f5-f7d7-4ac2-a15a-2b16025a00b4" containerName="pull" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.078448 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="51fa29f5-f7d7-4ac2-a15a-2b16025a00b4" containerName="pull" Dec 06 14:08:17 crc kubenswrapper[4706]: E1206 14:08:17.078473 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51fa29f5-f7d7-4ac2-a15a-2b16025a00b4" containerName="extract" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.078486 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="51fa29f5-f7d7-4ac2-a15a-2b16025a00b4" containerName="extract" Dec 06 14:08:17 crc kubenswrapper[4706]: E1206 14:08:17.078504 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51fa29f5-f7d7-4ac2-a15a-2b16025a00b4" containerName="util" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.078517 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="51fa29f5-f7d7-4ac2-a15a-2b16025a00b4" containerName="util" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.078695 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="51fa29f5-f7d7-4ac2-a15a-2b16025a00b4" containerName="extract" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.079355 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.083566 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.083808 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-sftpk" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.083873 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.086196 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.089357 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.093345 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9"] Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.183676 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsbvj\" (UniqueName: \"kubernetes.io/projected/5acf3c41-82f9-4b93-8742-1b4975610426-kube-api-access-qsbvj\") pod \"metallb-operator-controller-manager-548bff8fd9-mt2r9\" (UID: \"5acf3c41-82f9-4b93-8742-1b4975610426\") " pod="metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.183874 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5acf3c41-82f9-4b93-8742-1b4975610426-webhook-cert\") pod \"metallb-operator-controller-manager-548bff8fd9-mt2r9\" (UID: \"5acf3c41-82f9-4b93-8742-1b4975610426\") " pod="metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.183907 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5acf3c41-82f9-4b93-8742-1b4975610426-apiservice-cert\") pod \"metallb-operator-controller-manager-548bff8fd9-mt2r9\" (UID: \"5acf3c41-82f9-4b93-8742-1b4975610426\") " pod="metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.285629 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5acf3c41-82f9-4b93-8742-1b4975610426-webhook-cert\") pod \"metallb-operator-controller-manager-548bff8fd9-mt2r9\" (UID: \"5acf3c41-82f9-4b93-8742-1b4975610426\") " pod="metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.285686 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5acf3c41-82f9-4b93-8742-1b4975610426-apiservice-cert\") pod \"metallb-operator-controller-manager-548bff8fd9-mt2r9\" (UID: \"5acf3c41-82f9-4b93-8742-1b4975610426\") " pod="metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.285764 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsbvj\" (UniqueName: \"kubernetes.io/projected/5acf3c41-82f9-4b93-8742-1b4975610426-kube-api-access-qsbvj\") pod \"metallb-operator-controller-manager-548bff8fd9-mt2r9\" (UID: \"5acf3c41-82f9-4b93-8742-1b4975610426\") " pod="metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.293379 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5acf3c41-82f9-4b93-8742-1b4975610426-webhook-cert\") pod \"metallb-operator-controller-manager-548bff8fd9-mt2r9\" (UID: \"5acf3c41-82f9-4b93-8742-1b4975610426\") " pod="metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.294659 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5acf3c41-82f9-4b93-8742-1b4975610426-apiservice-cert\") pod \"metallb-operator-controller-manager-548bff8fd9-mt2r9\" (UID: \"5acf3c41-82f9-4b93-8742-1b4975610426\") " pod="metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.302573 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsbvj\" (UniqueName: \"kubernetes.io/projected/5acf3c41-82f9-4b93-8742-1b4975610426-kube-api-access-qsbvj\") pod \"metallb-operator-controller-manager-548bff8fd9-mt2r9\" (UID: \"5acf3c41-82f9-4b93-8742-1b4975610426\") " pod="metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.344723 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm"] Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.346165 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.347924 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.348378 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-8rc4q" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.350537 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.361632 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm"] Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.400560 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.488217 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d16f9b04-736c-4331-8ad8-58d14bd858e6-webhook-cert\") pod \"metallb-operator-webhook-server-56bc6d5c44-tg7hm\" (UID: \"d16f9b04-736c-4331-8ad8-58d14bd858e6\") " pod="metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.488296 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d16f9b04-736c-4331-8ad8-58d14bd858e6-apiservice-cert\") pod \"metallb-operator-webhook-server-56bc6d5c44-tg7hm\" (UID: \"d16f9b04-736c-4331-8ad8-58d14bd858e6\") " pod="metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.488338 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm8gq\" (UniqueName: \"kubernetes.io/projected/d16f9b04-736c-4331-8ad8-58d14bd858e6-kube-api-access-tm8gq\") pod \"metallb-operator-webhook-server-56bc6d5c44-tg7hm\" (UID: \"d16f9b04-736c-4331-8ad8-58d14bd858e6\") " pod="metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.589168 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d16f9b04-736c-4331-8ad8-58d14bd858e6-apiservice-cert\") pod \"metallb-operator-webhook-server-56bc6d5c44-tg7hm\" (UID: \"d16f9b04-736c-4331-8ad8-58d14bd858e6\") " pod="metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.589230 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm8gq\" (UniqueName: \"kubernetes.io/projected/d16f9b04-736c-4331-8ad8-58d14bd858e6-kube-api-access-tm8gq\") pod \"metallb-operator-webhook-server-56bc6d5c44-tg7hm\" (UID: \"d16f9b04-736c-4331-8ad8-58d14bd858e6\") " pod="metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.589294 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d16f9b04-736c-4331-8ad8-58d14bd858e6-webhook-cert\") pod \"metallb-operator-webhook-server-56bc6d5c44-tg7hm\" (UID: \"d16f9b04-736c-4331-8ad8-58d14bd858e6\") " pod="metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.595694 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d16f9b04-736c-4331-8ad8-58d14bd858e6-webhook-cert\") pod \"metallb-operator-webhook-server-56bc6d5c44-tg7hm\" (UID: \"d16f9b04-736c-4331-8ad8-58d14bd858e6\") " pod="metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.604003 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm8gq\" (UniqueName: \"kubernetes.io/projected/d16f9b04-736c-4331-8ad8-58d14bd858e6-kube-api-access-tm8gq\") pod \"metallb-operator-webhook-server-56bc6d5c44-tg7hm\" (UID: \"d16f9b04-736c-4331-8ad8-58d14bd858e6\") " pod="metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.612791 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d16f9b04-736c-4331-8ad8-58d14bd858e6-apiservice-cert\") pod \"metallb-operator-webhook-server-56bc6d5c44-tg7hm\" (UID: \"d16f9b04-736c-4331-8ad8-58d14bd858e6\") " pod="metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.659621 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm" Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.698064 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9"] Dec 06 14:08:17 crc kubenswrapper[4706]: W1206 14:08:17.714070 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5acf3c41_82f9_4b93_8742_1b4975610426.slice/crio-67f4be7b84406b826b768c3a54f39789e44264dd4e3416aec69ba61846314ebf WatchSource:0}: Error finding container 67f4be7b84406b826b768c3a54f39789e44264dd4e3416aec69ba61846314ebf: Status 404 returned error can't find the container with id 67f4be7b84406b826b768c3a54f39789e44264dd4e3416aec69ba61846314ebf Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.756783 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9" event={"ID":"5acf3c41-82f9-4b93-8742-1b4975610426","Type":"ContainerStarted","Data":"67f4be7b84406b826b768c3a54f39789e44264dd4e3416aec69ba61846314ebf"} Dec 06 14:08:17 crc kubenswrapper[4706]: I1206 14:08:17.873017 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm"] Dec 06 14:08:17 crc kubenswrapper[4706]: W1206 14:08:17.881462 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd16f9b04_736c_4331_8ad8_58d14bd858e6.slice/crio-86d3ebe00f80ab5e197c9c6e81ca23b39168dfa73bc162928057cea35baa4958 WatchSource:0}: Error finding container 86d3ebe00f80ab5e197c9c6e81ca23b39168dfa73bc162928057cea35baa4958: Status 404 returned error can't find the container with id 86d3ebe00f80ab5e197c9c6e81ca23b39168dfa73bc162928057cea35baa4958 Dec 06 14:08:18 crc kubenswrapper[4706]: I1206 14:08:18.768924 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm" event={"ID":"d16f9b04-736c-4331-8ad8-58d14bd858e6","Type":"ContainerStarted","Data":"86d3ebe00f80ab5e197c9c6e81ca23b39168dfa73bc162928057cea35baa4958"} Dec 06 14:08:20 crc kubenswrapper[4706]: I1206 14:08:20.781641 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9" event={"ID":"5acf3c41-82f9-4b93-8742-1b4975610426","Type":"ContainerStarted","Data":"d0777f414138f13e2f1232a87bda1849dfa0ad170cd2592ce3c021fdcfb979b5"} Dec 06 14:08:20 crc kubenswrapper[4706]: I1206 14:08:20.782203 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9" Dec 06 14:08:20 crc kubenswrapper[4706]: I1206 14:08:20.808390 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9" podStartSLOduration=0.937118834 podStartE2EDuration="3.808323541s" podCreationTimestamp="2025-12-06 14:08:17 +0000 UTC" firstStartedPulling="2025-12-06 14:08:17.716574297 +0000 UTC m=+740.552367154" lastFinishedPulling="2025-12-06 14:08:20.587779004 +0000 UTC m=+743.423571861" observedRunningTime="2025-12-06 14:08:20.805970108 +0000 UTC m=+743.641762965" watchObservedRunningTime="2025-12-06 14:08:20.808323541 +0000 UTC m=+743.644116408" Dec 06 14:08:22 crc kubenswrapper[4706]: I1206 14:08:22.794494 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm" event={"ID":"d16f9b04-736c-4331-8ad8-58d14bd858e6","Type":"ContainerStarted","Data":"497321f743a607a61d98fae4df069c0e7322e264797d701db02eacc2043ef10a"} Dec 06 14:08:22 crc kubenswrapper[4706]: I1206 14:08:22.795004 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm" Dec 06 14:08:22 crc kubenswrapper[4706]: I1206 14:08:22.827616 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm" podStartSLOduration=1.147036728 podStartE2EDuration="5.827596073s" podCreationTimestamp="2025-12-06 14:08:17 +0000 UTC" firstStartedPulling="2025-12-06 14:08:17.889388248 +0000 UTC m=+740.725181105" lastFinishedPulling="2025-12-06 14:08:22.569947593 +0000 UTC m=+745.405740450" observedRunningTime="2025-12-06 14:08:22.823968357 +0000 UTC m=+745.659761214" watchObservedRunningTime="2025-12-06 14:08:22.827596073 +0000 UTC m=+745.663388940" Dec 06 14:08:36 crc kubenswrapper[4706]: I1206 14:08:36.321696 4706 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 14:08:37 crc kubenswrapper[4706]: I1206 14:08:37.663887 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-56bc6d5c44-tg7hm" Dec 06 14:08:44 crc kubenswrapper[4706]: I1206 14:08:44.580036 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:08:44 crc kubenswrapper[4706]: I1206 14:08:44.580713 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:08:57 crc kubenswrapper[4706]: I1206 14:08:57.403860 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-548bff8fd9-mt2r9" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.231100 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-5vmhj"] Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.232226 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5vmhj" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.234573 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-fk67w" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.234813 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.240119 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-zffsv"] Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.242905 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.246244 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.246579 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.255052 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-5vmhj"] Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.331372 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-w6jgb"] Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.332612 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-w6jgb" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.334829 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.335442 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.335675 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.335924 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-rjmm4" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.364790 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-bcgwr"] Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.374582 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-bcgwr" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.378183 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.396759 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-frr-conf\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.396800 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-frr-sockets\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.396829 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5shqc\" (UniqueName: \"kubernetes.io/projected/7cd749c9-c857-4019-9fb3-f2c53f0b9be8-kube-api-access-5shqc\") pod \"frr-k8s-webhook-server-7fcb986d4-5vmhj\" (UID: \"7cd749c9-c857-4019-9fb3-f2c53f0b9be8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5vmhj" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.396854 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-metrics\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.396875 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-reloader\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.396901 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7cd749c9-c857-4019-9fb3-f2c53f0b9be8-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-5vmhj\" (UID: \"7cd749c9-c857-4019-9fb3-f2c53f0b9be8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5vmhj" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.396930 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-frr-startup\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.396946 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx9s2\" (UniqueName: \"kubernetes.io/projected/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-kube-api-access-mx9s2\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.396960 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-metrics-certs\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.397433 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-bcgwr"] Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.498538 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5shqc\" (UniqueName: \"kubernetes.io/projected/7cd749c9-c857-4019-9fb3-f2c53f0b9be8-kube-api-access-5shqc\") pod \"frr-k8s-webhook-server-7fcb986d4-5vmhj\" (UID: \"7cd749c9-c857-4019-9fb3-f2c53f0b9be8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5vmhj" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.498607 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99nrx\" (UniqueName: \"kubernetes.io/projected/d6c3a1b0-6512-40a8-aac5-bc6629793ab8-kube-api-access-99nrx\") pod \"controller-f8648f98b-bcgwr\" (UID: \"d6c3a1b0-6512-40a8-aac5-bc6629793ab8\") " pod="metallb-system/controller-f8648f98b-bcgwr" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.498638 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wdx4\" (UniqueName: \"kubernetes.io/projected/afd1dfce-966c-4506-bf87-3e2730857136-kube-api-access-5wdx4\") pod \"speaker-w6jgb\" (UID: \"afd1dfce-966c-4506-bf87-3e2730857136\") " pod="metallb-system/speaker-w6jgb" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.498705 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-metrics\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.498745 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d6c3a1b0-6512-40a8-aac5-bc6629793ab8-metrics-certs\") pod \"controller-f8648f98b-bcgwr\" (UID: \"d6c3a1b0-6512-40a8-aac5-bc6629793ab8\") " pod="metallb-system/controller-f8648f98b-bcgwr" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.498767 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-reloader\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.498784 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d6c3a1b0-6512-40a8-aac5-bc6629793ab8-cert\") pod \"controller-f8648f98b-bcgwr\" (UID: \"d6c3a1b0-6512-40a8-aac5-bc6629793ab8\") " pod="metallb-system/controller-f8648f98b-bcgwr" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.498804 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/afd1dfce-966c-4506-bf87-3e2730857136-memberlist\") pod \"speaker-w6jgb\" (UID: \"afd1dfce-966c-4506-bf87-3e2730857136\") " pod="metallb-system/speaker-w6jgb" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.498831 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7cd749c9-c857-4019-9fb3-f2c53f0b9be8-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-5vmhj\" (UID: \"7cd749c9-c857-4019-9fb3-f2c53f0b9be8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5vmhj" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.498847 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/afd1dfce-966c-4506-bf87-3e2730857136-metrics-certs\") pod \"speaker-w6jgb\" (UID: \"afd1dfce-966c-4506-bf87-3e2730857136\") " pod="metallb-system/speaker-w6jgb" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.498868 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/afd1dfce-966c-4506-bf87-3e2730857136-metallb-excludel2\") pod \"speaker-w6jgb\" (UID: \"afd1dfce-966c-4506-bf87-3e2730857136\") " pod="metallb-system/speaker-w6jgb" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.498893 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-frr-startup\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.498909 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx9s2\" (UniqueName: \"kubernetes.io/projected/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-kube-api-access-mx9s2\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.498926 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-metrics-certs\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: E1206 14:08:58.499020 4706 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.499029 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-frr-conf\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: E1206 14:08:58.499099 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cd749c9-c857-4019-9fb3-f2c53f0b9be8-cert podName:7cd749c9-c857-4019-9fb3-f2c53f0b9be8 nodeName:}" failed. No retries permitted until 2025-12-06 14:08:58.999069468 +0000 UTC m=+781.834862335 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7cd749c9-c857-4019-9fb3-f2c53f0b9be8-cert") pod "frr-k8s-webhook-server-7fcb986d4-5vmhj" (UID: "7cd749c9-c857-4019-9fb3-f2c53f0b9be8") : secret "frr-k8s-webhook-server-cert" not found Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.499160 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-frr-sockets\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.499449 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-reloader\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.499470 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-frr-conf\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.499703 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-frr-sockets\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.500100 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-frr-startup\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.500248 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-metrics\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.504767 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-metrics-certs\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.513148 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5shqc\" (UniqueName: \"kubernetes.io/projected/7cd749c9-c857-4019-9fb3-f2c53f0b9be8-kube-api-access-5shqc\") pod \"frr-k8s-webhook-server-7fcb986d4-5vmhj\" (UID: \"7cd749c9-c857-4019-9fb3-f2c53f0b9be8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5vmhj" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.519162 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx9s2\" (UniqueName: \"kubernetes.io/projected/cb8721ae-d9da-4caf-87d3-f2dbf09c037b-kube-api-access-mx9s2\") pod \"frr-k8s-zffsv\" (UID: \"cb8721ae-d9da-4caf-87d3-f2dbf09c037b\") " pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.566299 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zffsv" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.600001 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99nrx\" (UniqueName: \"kubernetes.io/projected/d6c3a1b0-6512-40a8-aac5-bc6629793ab8-kube-api-access-99nrx\") pod \"controller-f8648f98b-bcgwr\" (UID: \"d6c3a1b0-6512-40a8-aac5-bc6629793ab8\") " pod="metallb-system/controller-f8648f98b-bcgwr" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.600040 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wdx4\" (UniqueName: \"kubernetes.io/projected/afd1dfce-966c-4506-bf87-3e2730857136-kube-api-access-5wdx4\") pod \"speaker-w6jgb\" (UID: \"afd1dfce-966c-4506-bf87-3e2730857136\") " pod="metallb-system/speaker-w6jgb" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.600064 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d6c3a1b0-6512-40a8-aac5-bc6629793ab8-metrics-certs\") pod \"controller-f8648f98b-bcgwr\" (UID: \"d6c3a1b0-6512-40a8-aac5-bc6629793ab8\") " pod="metallb-system/controller-f8648f98b-bcgwr" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.600100 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d6c3a1b0-6512-40a8-aac5-bc6629793ab8-cert\") pod \"controller-f8648f98b-bcgwr\" (UID: \"d6c3a1b0-6512-40a8-aac5-bc6629793ab8\") " pod="metallb-system/controller-f8648f98b-bcgwr" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.600116 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/afd1dfce-966c-4506-bf87-3e2730857136-memberlist\") pod \"speaker-w6jgb\" (UID: \"afd1dfce-966c-4506-bf87-3e2730857136\") " pod="metallb-system/speaker-w6jgb" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.600158 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/afd1dfce-966c-4506-bf87-3e2730857136-metrics-certs\") pod \"speaker-w6jgb\" (UID: \"afd1dfce-966c-4506-bf87-3e2730857136\") " pod="metallb-system/speaker-w6jgb" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.600177 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/afd1dfce-966c-4506-bf87-3e2730857136-metallb-excludel2\") pod \"speaker-w6jgb\" (UID: \"afd1dfce-966c-4506-bf87-3e2730857136\") " pod="metallb-system/speaker-w6jgb" Dec 06 14:08:58 crc kubenswrapper[4706]: E1206 14:08:58.601080 4706 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 14:08:58 crc kubenswrapper[4706]: E1206 14:08:58.601139 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/afd1dfce-966c-4506-bf87-3e2730857136-memberlist podName:afd1dfce-966c-4506-bf87-3e2730857136 nodeName:}" failed. No retries permitted until 2025-12-06 14:08:59.101110343 +0000 UTC m=+781.936903200 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/afd1dfce-966c-4506-bf87-3e2730857136-memberlist") pod "speaker-w6jgb" (UID: "afd1dfce-966c-4506-bf87-3e2730857136") : secret "metallb-memberlist" not found Dec 06 14:08:58 crc kubenswrapper[4706]: E1206 14:08:58.601597 4706 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 06 14:08:58 crc kubenswrapper[4706]: E1206 14:08:58.601628 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/afd1dfce-966c-4506-bf87-3e2730857136-metrics-certs podName:afd1dfce-966c-4506-bf87-3e2730857136 nodeName:}" failed. No retries permitted until 2025-12-06 14:08:59.101618966 +0000 UTC m=+781.937411823 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/afd1dfce-966c-4506-bf87-3e2730857136-metrics-certs") pod "speaker-w6jgb" (UID: "afd1dfce-966c-4506-bf87-3e2730857136") : secret "speaker-certs-secret" not found Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.602109 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/afd1dfce-966c-4506-bf87-3e2730857136-metallb-excludel2\") pod \"speaker-w6jgb\" (UID: \"afd1dfce-966c-4506-bf87-3e2730857136\") " pod="metallb-system/speaker-w6jgb" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.604396 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.604821 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d6c3a1b0-6512-40a8-aac5-bc6629793ab8-metrics-certs\") pod \"controller-f8648f98b-bcgwr\" (UID: \"d6c3a1b0-6512-40a8-aac5-bc6629793ab8\") " pod="metallb-system/controller-f8648f98b-bcgwr" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.613435 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d6c3a1b0-6512-40a8-aac5-bc6629793ab8-cert\") pod \"controller-f8648f98b-bcgwr\" (UID: \"d6c3a1b0-6512-40a8-aac5-bc6629793ab8\") " pod="metallb-system/controller-f8648f98b-bcgwr" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.618260 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wdx4\" (UniqueName: \"kubernetes.io/projected/afd1dfce-966c-4506-bf87-3e2730857136-kube-api-access-5wdx4\") pod \"speaker-w6jgb\" (UID: \"afd1dfce-966c-4506-bf87-3e2730857136\") " pod="metallb-system/speaker-w6jgb" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.627673 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99nrx\" (UniqueName: \"kubernetes.io/projected/d6c3a1b0-6512-40a8-aac5-bc6629793ab8-kube-api-access-99nrx\") pod \"controller-f8648f98b-bcgwr\" (UID: \"d6c3a1b0-6512-40a8-aac5-bc6629793ab8\") " pod="metallb-system/controller-f8648f98b-bcgwr" Dec 06 14:08:58 crc kubenswrapper[4706]: I1206 14:08:58.694529 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-bcgwr" Dec 06 14:08:59 crc kubenswrapper[4706]: I1206 14:08:59.022379 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7cd749c9-c857-4019-9fb3-f2c53f0b9be8-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-5vmhj\" (UID: \"7cd749c9-c857-4019-9fb3-f2c53f0b9be8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5vmhj" Dec 06 14:08:59 crc kubenswrapper[4706]: I1206 14:08:59.028794 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7cd749c9-c857-4019-9fb3-f2c53f0b9be8-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-5vmhj\" (UID: \"7cd749c9-c857-4019-9fb3-f2c53f0b9be8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5vmhj" Dec 06 14:08:59 crc kubenswrapper[4706]: I1206 14:08:59.034553 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zffsv" event={"ID":"cb8721ae-d9da-4caf-87d3-f2dbf09c037b","Type":"ContainerStarted","Data":"e24ccfbd60a84244bb9a6bda15e1e51bf233e5c9919a701aae5c5c7c6f3ca566"} Dec 06 14:08:59 crc kubenswrapper[4706]: I1206 14:08:59.124565 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/afd1dfce-966c-4506-bf87-3e2730857136-metrics-certs\") pod \"speaker-w6jgb\" (UID: \"afd1dfce-966c-4506-bf87-3e2730857136\") " pod="metallb-system/speaker-w6jgb" Dec 06 14:08:59 crc kubenswrapper[4706]: I1206 14:08:59.124736 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/afd1dfce-966c-4506-bf87-3e2730857136-memberlist\") pod \"speaker-w6jgb\" (UID: \"afd1dfce-966c-4506-bf87-3e2730857136\") " pod="metallb-system/speaker-w6jgb" Dec 06 14:08:59 crc kubenswrapper[4706]: E1206 14:08:59.124836 4706 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 14:08:59 crc kubenswrapper[4706]: E1206 14:08:59.124922 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/afd1dfce-966c-4506-bf87-3e2730857136-memberlist podName:afd1dfce-966c-4506-bf87-3e2730857136 nodeName:}" failed. No retries permitted until 2025-12-06 14:09:00.124899249 +0000 UTC m=+782.960692126 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/afd1dfce-966c-4506-bf87-3e2730857136-memberlist") pod "speaker-w6jgb" (UID: "afd1dfce-966c-4506-bf87-3e2730857136") : secret "metallb-memberlist" not found Dec 06 14:08:59 crc kubenswrapper[4706]: I1206 14:08:59.129715 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/afd1dfce-966c-4506-bf87-3e2730857136-metrics-certs\") pod \"speaker-w6jgb\" (UID: \"afd1dfce-966c-4506-bf87-3e2730857136\") " pod="metallb-system/speaker-w6jgb" Dec 06 14:08:59 crc kubenswrapper[4706]: I1206 14:08:59.136079 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-bcgwr"] Dec 06 14:08:59 crc kubenswrapper[4706]: W1206 14:08:59.145113 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6c3a1b0_6512_40a8_aac5_bc6629793ab8.slice/crio-95bf56359edc3ebb73ef83d70f61a417d1f6ae8468ad14dae945c32800645521 WatchSource:0}: Error finding container 95bf56359edc3ebb73ef83d70f61a417d1f6ae8468ad14dae945c32800645521: Status 404 returned error can't find the container with id 95bf56359edc3ebb73ef83d70f61a417d1f6ae8468ad14dae945c32800645521 Dec 06 14:08:59 crc kubenswrapper[4706]: I1206 14:08:59.155882 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5vmhj" Dec 06 14:08:59 crc kubenswrapper[4706]: I1206 14:08:59.597794 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-5vmhj"] Dec 06 14:08:59 crc kubenswrapper[4706]: W1206 14:08:59.607045 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7cd749c9_c857_4019_9fb3_f2c53f0b9be8.slice/crio-dd507f38cd28d4d6d6c0ecb62e46acec1ff61b00fd6c4ba45dfd3e809d5281d3 WatchSource:0}: Error finding container dd507f38cd28d4d6d6c0ecb62e46acec1ff61b00fd6c4ba45dfd3e809d5281d3: Status 404 returned error can't find the container with id dd507f38cd28d4d6d6c0ecb62e46acec1ff61b00fd6c4ba45dfd3e809d5281d3 Dec 06 14:09:00 crc kubenswrapper[4706]: I1206 14:09:00.041811 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-bcgwr" event={"ID":"d6c3a1b0-6512-40a8-aac5-bc6629793ab8","Type":"ContainerStarted","Data":"cfd17f864c3a91b9bb354ad0f9faf63478f8800ba5cf81bd19215ad64890e04c"} Dec 06 14:09:00 crc kubenswrapper[4706]: I1206 14:09:00.042385 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-bcgwr" event={"ID":"d6c3a1b0-6512-40a8-aac5-bc6629793ab8","Type":"ContainerStarted","Data":"95bf56359edc3ebb73ef83d70f61a417d1f6ae8468ad14dae945c32800645521"} Dec 06 14:09:00 crc kubenswrapper[4706]: I1206 14:09:00.042594 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5vmhj" event={"ID":"7cd749c9-c857-4019-9fb3-f2c53f0b9be8","Type":"ContainerStarted","Data":"dd507f38cd28d4d6d6c0ecb62e46acec1ff61b00fd6c4ba45dfd3e809d5281d3"} Dec 06 14:09:00 crc kubenswrapper[4706]: I1206 14:09:00.136494 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/afd1dfce-966c-4506-bf87-3e2730857136-memberlist\") pod \"speaker-w6jgb\" (UID: \"afd1dfce-966c-4506-bf87-3e2730857136\") " pod="metallb-system/speaker-w6jgb" Dec 06 14:09:00 crc kubenswrapper[4706]: I1206 14:09:00.156719 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/afd1dfce-966c-4506-bf87-3e2730857136-memberlist\") pod \"speaker-w6jgb\" (UID: \"afd1dfce-966c-4506-bf87-3e2730857136\") " pod="metallb-system/speaker-w6jgb" Dec 06 14:09:00 crc kubenswrapper[4706]: I1206 14:09:00.446194 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-w6jgb" Dec 06 14:09:00 crc kubenswrapper[4706]: W1206 14:09:00.483641 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafd1dfce_966c_4506_bf87_3e2730857136.slice/crio-a31789e9d2a040465a0d16555d1da3f5bcb9565a83e0112733ba0e155ac23ccf WatchSource:0}: Error finding container a31789e9d2a040465a0d16555d1da3f5bcb9565a83e0112733ba0e155ac23ccf: Status 404 returned error can't find the container with id a31789e9d2a040465a0d16555d1da3f5bcb9565a83e0112733ba0e155ac23ccf Dec 06 14:09:01 crc kubenswrapper[4706]: I1206 14:09:01.051926 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-w6jgb" event={"ID":"afd1dfce-966c-4506-bf87-3e2730857136","Type":"ContainerStarted","Data":"e089ef5b7478ac5494adfd06619de8b150fa82edf3798e518c5315d6258d2d14"} Dec 06 14:09:01 crc kubenswrapper[4706]: I1206 14:09:01.052186 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-w6jgb" event={"ID":"afd1dfce-966c-4506-bf87-3e2730857136","Type":"ContainerStarted","Data":"a31789e9d2a040465a0d16555d1da3f5bcb9565a83e0112733ba0e155ac23ccf"} Dec 06 14:09:03 crc kubenswrapper[4706]: I1206 14:09:03.067226 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-bcgwr" event={"ID":"d6c3a1b0-6512-40a8-aac5-bc6629793ab8","Type":"ContainerStarted","Data":"4fa72e89963c5bbd93407911db188a755ddd3350eadda462aa22c9ac98a95c92"} Dec 06 14:09:03 crc kubenswrapper[4706]: I1206 14:09:03.068958 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-bcgwr" Dec 06 14:09:03 crc kubenswrapper[4706]: I1206 14:09:03.069670 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-w6jgb" event={"ID":"afd1dfce-966c-4506-bf87-3e2730857136","Type":"ContainerStarted","Data":"b3d99f58b492c4f6ab6afc4b2fa17edc373505b15007b17e11745d6ffb51f254"} Dec 06 14:09:03 crc kubenswrapper[4706]: I1206 14:09:03.069816 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-w6jgb" Dec 06 14:09:03 crc kubenswrapper[4706]: I1206 14:09:03.090391 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-bcgwr" podStartSLOduration=2.146252566 podStartE2EDuration="5.090370426s" podCreationTimestamp="2025-12-06 14:08:58 +0000 UTC" firstStartedPulling="2025-12-06 14:08:59.354483926 +0000 UTC m=+782.190276773" lastFinishedPulling="2025-12-06 14:09:02.298601756 +0000 UTC m=+785.134394633" observedRunningTime="2025-12-06 14:09:03.088409694 +0000 UTC m=+785.924202571" watchObservedRunningTime="2025-12-06 14:09:03.090370426 +0000 UTC m=+785.926163283" Dec 06 14:09:03 crc kubenswrapper[4706]: I1206 14:09:03.108995 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-w6jgb" podStartSLOduration=3.555519176 podStartE2EDuration="5.10897574s" podCreationTimestamp="2025-12-06 14:08:58 +0000 UTC" firstStartedPulling="2025-12-06 14:09:00.737026037 +0000 UTC m=+783.572818894" lastFinishedPulling="2025-12-06 14:09:02.290482601 +0000 UTC m=+785.126275458" observedRunningTime="2025-12-06 14:09:03.104638125 +0000 UTC m=+785.940430992" watchObservedRunningTime="2025-12-06 14:09:03.10897574 +0000 UTC m=+785.944768597" Dec 06 14:09:06 crc kubenswrapper[4706]: I1206 14:09:06.100571 4706 generic.go:334] "Generic (PLEG): container finished" podID="cb8721ae-d9da-4caf-87d3-f2dbf09c037b" containerID="24cf3beb5632b32fd8004606894c2f6c5ab6bac9b0ba6c427842d9b257d2fbe1" exitCode=0 Dec 06 14:09:06 crc kubenswrapper[4706]: I1206 14:09:06.100653 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zffsv" event={"ID":"cb8721ae-d9da-4caf-87d3-f2dbf09c037b","Type":"ContainerDied","Data":"24cf3beb5632b32fd8004606894c2f6c5ab6bac9b0ba6c427842d9b257d2fbe1"} Dec 06 14:09:06 crc kubenswrapper[4706]: I1206 14:09:06.107103 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5vmhj" event={"ID":"7cd749c9-c857-4019-9fb3-f2c53f0b9be8","Type":"ContainerStarted","Data":"5082a0d46a14f4d3492c2ffd91e18b7de7ecfb8f6c1ef391cb3af0d00c7b6cf1"} Dec 06 14:09:06 crc kubenswrapper[4706]: I1206 14:09:06.107293 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5vmhj" Dec 06 14:09:07 crc kubenswrapper[4706]: I1206 14:09:07.119050 4706 generic.go:334] "Generic (PLEG): container finished" podID="cb8721ae-d9da-4caf-87d3-f2dbf09c037b" containerID="df32d34225cdcdc3100c2220f20ec6df8e7b048d4a380933f1482177f83956c8" exitCode=0 Dec 06 14:09:07 crc kubenswrapper[4706]: I1206 14:09:07.119112 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zffsv" event={"ID":"cb8721ae-d9da-4caf-87d3-f2dbf09c037b","Type":"ContainerDied","Data":"df32d34225cdcdc3100c2220f20ec6df8e7b048d4a380933f1482177f83956c8"} Dec 06 14:09:07 crc kubenswrapper[4706]: I1206 14:09:07.161504 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5vmhj" podStartSLOduration=3.057625809 podStartE2EDuration="9.161474645s" podCreationTimestamp="2025-12-06 14:08:58 +0000 UTC" firstStartedPulling="2025-12-06 14:08:59.609037424 +0000 UTC m=+782.444830291" lastFinishedPulling="2025-12-06 14:09:05.71288627 +0000 UTC m=+788.548679127" observedRunningTime="2025-12-06 14:09:06.167047311 +0000 UTC m=+789.002840208" watchObservedRunningTime="2025-12-06 14:09:07.161474645 +0000 UTC m=+789.997267542" Dec 06 14:09:07 crc kubenswrapper[4706]: E1206 14:09:07.463114 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb8721ae_d9da_4caf_87d3_f2dbf09c037b.slice/crio-conmon-a2a16adaca0ea3b286e93e798f80dbfcb96003378e8253e2b97d933bda379dd4.scope\": RecentStats: unable to find data in memory cache]" Dec 06 14:09:08 crc kubenswrapper[4706]: I1206 14:09:08.129200 4706 generic.go:334] "Generic (PLEG): container finished" podID="cb8721ae-d9da-4caf-87d3-f2dbf09c037b" containerID="a2a16adaca0ea3b286e93e798f80dbfcb96003378e8253e2b97d933bda379dd4" exitCode=0 Dec 06 14:09:08 crc kubenswrapper[4706]: I1206 14:09:08.129275 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zffsv" event={"ID":"cb8721ae-d9da-4caf-87d3-f2dbf09c037b","Type":"ContainerDied","Data":"a2a16adaca0ea3b286e93e798f80dbfcb96003378e8253e2b97d933bda379dd4"} Dec 06 14:09:09 crc kubenswrapper[4706]: I1206 14:09:09.150990 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zffsv" event={"ID":"cb8721ae-d9da-4caf-87d3-f2dbf09c037b","Type":"ContainerStarted","Data":"3335697efc355b64b687ced6eaf969cf3c6d34c3d5912224c0b9bb0f95d450ec"} Dec 06 14:09:09 crc kubenswrapper[4706]: I1206 14:09:09.151345 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zffsv" event={"ID":"cb8721ae-d9da-4caf-87d3-f2dbf09c037b","Type":"ContainerStarted","Data":"b70e9fb3deb0fb3c6d404414f1caa48f19dc393c8c03783ef998859f348ce16c"} Dec 06 14:09:09 crc kubenswrapper[4706]: I1206 14:09:09.151357 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zffsv" event={"ID":"cb8721ae-d9da-4caf-87d3-f2dbf09c037b","Type":"ContainerStarted","Data":"c3576ace9268a8d0449489cded80083a03c851c74f177ce04cd0549429be88a6"} Dec 06 14:09:09 crc kubenswrapper[4706]: I1206 14:09:09.151366 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zffsv" event={"ID":"cb8721ae-d9da-4caf-87d3-f2dbf09c037b","Type":"ContainerStarted","Data":"c789789520f6eb9c53dfea551acdeb5d2b37815bca07c867841020318884a32e"} Dec 06 14:09:09 crc kubenswrapper[4706]: I1206 14:09:09.151376 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zffsv" event={"ID":"cb8721ae-d9da-4caf-87d3-f2dbf09c037b","Type":"ContainerStarted","Data":"df007b6fb5d87e9a789c6fee5b48963dd4a56a398db5c0bdd4de3574891a0dbd"} Dec 06 14:09:10 crc kubenswrapper[4706]: I1206 14:09:10.166915 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zffsv" event={"ID":"cb8721ae-d9da-4caf-87d3-f2dbf09c037b","Type":"ContainerStarted","Data":"e9670055287d40e73fa077a77b8ebbb8adffad9a30e0a6cc8d8ccc7dd498277b"} Dec 06 14:09:10 crc kubenswrapper[4706]: I1206 14:09:10.167532 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-zffsv" Dec 06 14:09:10 crc kubenswrapper[4706]: I1206 14:09:10.197692 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-zffsv" podStartSLOduration=5.196954355 podStartE2EDuration="12.197665738s" podCreationTimestamp="2025-12-06 14:08:58 +0000 UTC" firstStartedPulling="2025-12-06 14:08:58.712841375 +0000 UTC m=+781.548634242" lastFinishedPulling="2025-12-06 14:09:05.713552758 +0000 UTC m=+788.549345625" observedRunningTime="2025-12-06 14:09:10.19512328 +0000 UTC m=+793.030916157" watchObservedRunningTime="2025-12-06 14:09:10.197665738 +0000 UTC m=+793.033458605" Dec 06 14:09:10 crc kubenswrapper[4706]: I1206 14:09:10.453269 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-w6jgb" Dec 06 14:09:13 crc kubenswrapper[4706]: I1206 14:09:13.567514 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-zffsv" Dec 06 14:09:13 crc kubenswrapper[4706]: I1206 14:09:13.633136 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-zffsv" Dec 06 14:09:14 crc kubenswrapper[4706]: I1206 14:09:14.580117 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:09:14 crc kubenswrapper[4706]: I1206 14:09:14.580221 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:09:14 crc kubenswrapper[4706]: I1206 14:09:14.580303 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 14:09:14 crc kubenswrapper[4706]: I1206 14:09:14.581362 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8f8d8c75a654551bee4406e15ba9924159f4ad79dd336b58107e1b41c26cbce8"} pod="openshift-machine-config-operator/machine-config-daemon-67v56" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 14:09:14 crc kubenswrapper[4706]: I1206 14:09:14.581500 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" containerID="cri-o://8f8d8c75a654551bee4406e15ba9924159f4ad79dd336b58107e1b41c26cbce8" gracePeriod=600 Dec 06 14:09:16 crc kubenswrapper[4706]: I1206 14:09:16.215516 4706 generic.go:334] "Generic (PLEG): container finished" podID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerID="8f8d8c75a654551bee4406e15ba9924159f4ad79dd336b58107e1b41c26cbce8" exitCode=0 Dec 06 14:09:16 crc kubenswrapper[4706]: I1206 14:09:16.215610 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerDied","Data":"8f8d8c75a654551bee4406e15ba9924159f4ad79dd336b58107e1b41c26cbce8"} Dec 06 14:09:16 crc kubenswrapper[4706]: I1206 14:09:16.216402 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerStarted","Data":"fc4d148c623467502d8b194b1ce686e138f01f9dcba1a1b63282c7d88ff885f4"} Dec 06 14:09:16 crc kubenswrapper[4706]: I1206 14:09:16.216446 4706 scope.go:117] "RemoveContainer" containerID="22caa6fab03e2ecbf77dfdaba33f02fdf7629d6e7b5d96a389d97d1f79320ebe" Dec 06 14:09:16 crc kubenswrapper[4706]: I1206 14:09:16.538112 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-cq77l"] Dec 06 14:09:16 crc kubenswrapper[4706]: I1206 14:09:16.539279 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-cq77l" Dec 06 14:09:16 crc kubenswrapper[4706]: I1206 14:09:16.544231 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 06 14:09:16 crc kubenswrapper[4706]: I1206 14:09:16.544341 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-q5ggd" Dec 06 14:09:16 crc kubenswrapper[4706]: I1206 14:09:16.545077 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 06 14:09:16 crc kubenswrapper[4706]: I1206 14:09:16.549599 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-cq77l"] Dec 06 14:09:16 crc kubenswrapper[4706]: I1206 14:09:16.679155 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2j9r\" (UniqueName: \"kubernetes.io/projected/ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9-kube-api-access-v2j9r\") pod \"mariadb-operator-index-cq77l\" (UID: \"ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9\") " pod="openstack-operators/mariadb-operator-index-cq77l" Dec 06 14:09:16 crc kubenswrapper[4706]: I1206 14:09:16.780799 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2j9r\" (UniqueName: \"kubernetes.io/projected/ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9-kube-api-access-v2j9r\") pod \"mariadb-operator-index-cq77l\" (UID: \"ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9\") " pod="openstack-operators/mariadb-operator-index-cq77l" Dec 06 14:09:16 crc kubenswrapper[4706]: I1206 14:09:16.805701 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2j9r\" (UniqueName: \"kubernetes.io/projected/ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9-kube-api-access-v2j9r\") pod \"mariadb-operator-index-cq77l\" (UID: \"ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9\") " pod="openstack-operators/mariadb-operator-index-cq77l" Dec 06 14:09:16 crc kubenswrapper[4706]: I1206 14:09:16.863757 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-cq77l" Dec 06 14:09:17 crc kubenswrapper[4706]: I1206 14:09:17.102595 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-cq77l"] Dec 06 14:09:17 crc kubenswrapper[4706]: W1206 14:09:17.104572 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef80af1b_73e0_45f0_bcd9_25b8cf2d17f9.slice/crio-6b6cc2e086f2b7dd3dc846198136979fd5699d676f1330b19f03a840a927ab07 WatchSource:0}: Error finding container 6b6cc2e086f2b7dd3dc846198136979fd5699d676f1330b19f03a840a927ab07: Status 404 returned error can't find the container with id 6b6cc2e086f2b7dd3dc846198136979fd5699d676f1330b19f03a840a927ab07 Dec 06 14:09:17 crc kubenswrapper[4706]: I1206 14:09:17.226513 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-cq77l" event={"ID":"ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9","Type":"ContainerStarted","Data":"6b6cc2e086f2b7dd3dc846198136979fd5699d676f1330b19f03a840a927ab07"} Dec 06 14:09:18 crc kubenswrapper[4706]: I1206 14:09:18.233978 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-cq77l" event={"ID":"ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9","Type":"ContainerStarted","Data":"e80df8ebfe4ceeadfdb9dd4e004f4ceba231697b614b8ea13afe092239419dd1"} Dec 06 14:09:18 crc kubenswrapper[4706]: I1206 14:09:18.249933 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-cq77l" podStartSLOduration=1.434281255 podStartE2EDuration="2.249917967s" podCreationTimestamp="2025-12-06 14:09:16 +0000 UTC" firstStartedPulling="2025-12-06 14:09:17.107219804 +0000 UTC m=+799.943012661" lastFinishedPulling="2025-12-06 14:09:17.922856506 +0000 UTC m=+800.758649373" observedRunningTime="2025-12-06 14:09:18.247608806 +0000 UTC m=+801.083401663" watchObservedRunningTime="2025-12-06 14:09:18.249917967 +0000 UTC m=+801.085710824" Dec 06 14:09:18 crc kubenswrapper[4706]: I1206 14:09:18.585130 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-zffsv" Dec 06 14:09:18 crc kubenswrapper[4706]: I1206 14:09:18.701185 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-bcgwr" Dec 06 14:09:19 crc kubenswrapper[4706]: I1206 14:09:19.170052 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5vmhj" Dec 06 14:09:26 crc kubenswrapper[4706]: I1206 14:09:26.864381 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-cq77l" Dec 06 14:09:26 crc kubenswrapper[4706]: I1206 14:09:26.864988 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-cq77l" Dec 06 14:09:26 crc kubenswrapper[4706]: I1206 14:09:26.896608 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-cq77l" Dec 06 14:09:27 crc kubenswrapper[4706]: I1206 14:09:27.340576 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-cq77l" Dec 06 14:09:28 crc kubenswrapper[4706]: I1206 14:09:28.571041 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2"] Dec 06 14:09:28 crc kubenswrapper[4706]: I1206 14:09:28.573393 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" Dec 06 14:09:28 crc kubenswrapper[4706]: I1206 14:09:28.576111 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-fc6v2" Dec 06 14:09:28 crc kubenswrapper[4706]: I1206 14:09:28.589672 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2"] Dec 06 14:09:28 crc kubenswrapper[4706]: I1206 14:09:28.648647 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3e63f666-ace4-417b-ab05-50be328e8a64-bundle\") pod \"e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2\" (UID: \"3e63f666-ace4-417b-ab05-50be328e8a64\") " pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" Dec 06 14:09:28 crc kubenswrapper[4706]: I1206 14:09:28.648722 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-759k7\" (UniqueName: \"kubernetes.io/projected/3e63f666-ace4-417b-ab05-50be328e8a64-kube-api-access-759k7\") pod \"e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2\" (UID: \"3e63f666-ace4-417b-ab05-50be328e8a64\") " pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" Dec 06 14:09:28 crc kubenswrapper[4706]: I1206 14:09:28.648794 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3e63f666-ace4-417b-ab05-50be328e8a64-util\") pod \"e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2\" (UID: \"3e63f666-ace4-417b-ab05-50be328e8a64\") " pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" Dec 06 14:09:28 crc kubenswrapper[4706]: I1206 14:09:28.750131 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3e63f666-ace4-417b-ab05-50be328e8a64-bundle\") pod \"e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2\" (UID: \"3e63f666-ace4-417b-ab05-50be328e8a64\") " pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" Dec 06 14:09:28 crc kubenswrapper[4706]: I1206 14:09:28.750256 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-759k7\" (UniqueName: \"kubernetes.io/projected/3e63f666-ace4-417b-ab05-50be328e8a64-kube-api-access-759k7\") pod \"e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2\" (UID: \"3e63f666-ace4-417b-ab05-50be328e8a64\") " pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" Dec 06 14:09:28 crc kubenswrapper[4706]: I1206 14:09:28.750337 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3e63f666-ace4-417b-ab05-50be328e8a64-util\") pod \"e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2\" (UID: \"3e63f666-ace4-417b-ab05-50be328e8a64\") " pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" Dec 06 14:09:28 crc kubenswrapper[4706]: I1206 14:09:28.750872 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3e63f666-ace4-417b-ab05-50be328e8a64-bundle\") pod \"e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2\" (UID: \"3e63f666-ace4-417b-ab05-50be328e8a64\") " pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" Dec 06 14:09:28 crc kubenswrapper[4706]: I1206 14:09:28.751129 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3e63f666-ace4-417b-ab05-50be328e8a64-util\") pod \"e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2\" (UID: \"3e63f666-ace4-417b-ab05-50be328e8a64\") " pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" Dec 06 14:09:28 crc kubenswrapper[4706]: I1206 14:09:28.784936 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-759k7\" (UniqueName: \"kubernetes.io/projected/3e63f666-ace4-417b-ab05-50be328e8a64-kube-api-access-759k7\") pod \"e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2\" (UID: \"3e63f666-ace4-417b-ab05-50be328e8a64\") " pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" Dec 06 14:09:28 crc kubenswrapper[4706]: I1206 14:09:28.903370 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" Dec 06 14:09:29 crc kubenswrapper[4706]: I1206 14:09:29.175027 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2"] Dec 06 14:09:29 crc kubenswrapper[4706]: W1206 14:09:29.185505 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e63f666_ace4_417b_ab05_50be328e8a64.slice/crio-80c136375d81048ee2eb052060f18436929332bdb3b117ddc739a22d130d57d5 WatchSource:0}: Error finding container 80c136375d81048ee2eb052060f18436929332bdb3b117ddc739a22d130d57d5: Status 404 returned error can't find the container with id 80c136375d81048ee2eb052060f18436929332bdb3b117ddc739a22d130d57d5 Dec 06 14:09:29 crc kubenswrapper[4706]: I1206 14:09:29.318109 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" event={"ID":"3e63f666-ace4-417b-ab05-50be328e8a64","Type":"ContainerStarted","Data":"80c136375d81048ee2eb052060f18436929332bdb3b117ddc739a22d130d57d5"} Dec 06 14:09:30 crc kubenswrapper[4706]: I1206 14:09:30.329909 4706 generic.go:334] "Generic (PLEG): container finished" podID="3e63f666-ace4-417b-ab05-50be328e8a64" containerID="55e1d3d6f27102508e4473f17e2112843cd9400f821dcce0566586254bcec43b" exitCode=0 Dec 06 14:09:30 crc kubenswrapper[4706]: I1206 14:09:30.329969 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" event={"ID":"3e63f666-ace4-417b-ab05-50be328e8a64","Type":"ContainerDied","Data":"55e1d3d6f27102508e4473f17e2112843cd9400f821dcce0566586254bcec43b"} Dec 06 14:09:31 crc kubenswrapper[4706]: I1206 14:09:31.340367 4706 generic.go:334] "Generic (PLEG): container finished" podID="3e63f666-ace4-417b-ab05-50be328e8a64" containerID="c911be28060746c5240912071547f69459b978a79557e8ec1d05b255d87fd594" exitCode=0 Dec 06 14:09:31 crc kubenswrapper[4706]: I1206 14:09:31.340444 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" event={"ID":"3e63f666-ace4-417b-ab05-50be328e8a64","Type":"ContainerDied","Data":"c911be28060746c5240912071547f69459b978a79557e8ec1d05b255d87fd594"} Dec 06 14:09:32 crc kubenswrapper[4706]: I1206 14:09:32.353242 4706 generic.go:334] "Generic (PLEG): container finished" podID="3e63f666-ace4-417b-ab05-50be328e8a64" containerID="84be07ce6389a4820d53a56b02ac8426f433e23df7e3b1b177c0cd30560e5e81" exitCode=0 Dec 06 14:09:32 crc kubenswrapper[4706]: I1206 14:09:32.353357 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" event={"ID":"3e63f666-ace4-417b-ab05-50be328e8a64","Type":"ContainerDied","Data":"84be07ce6389a4820d53a56b02ac8426f433e23df7e3b1b177c0cd30560e5e81"} Dec 06 14:09:33 crc kubenswrapper[4706]: I1206 14:09:33.703540 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" Dec 06 14:09:33 crc kubenswrapper[4706]: I1206 14:09:33.828526 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-759k7\" (UniqueName: \"kubernetes.io/projected/3e63f666-ace4-417b-ab05-50be328e8a64-kube-api-access-759k7\") pod \"3e63f666-ace4-417b-ab05-50be328e8a64\" (UID: \"3e63f666-ace4-417b-ab05-50be328e8a64\") " Dec 06 14:09:33 crc kubenswrapper[4706]: I1206 14:09:33.828570 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3e63f666-ace4-417b-ab05-50be328e8a64-util\") pod \"3e63f666-ace4-417b-ab05-50be328e8a64\" (UID: \"3e63f666-ace4-417b-ab05-50be328e8a64\") " Dec 06 14:09:33 crc kubenswrapper[4706]: I1206 14:09:33.828605 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3e63f666-ace4-417b-ab05-50be328e8a64-bundle\") pod \"3e63f666-ace4-417b-ab05-50be328e8a64\" (UID: \"3e63f666-ace4-417b-ab05-50be328e8a64\") " Dec 06 14:09:33 crc kubenswrapper[4706]: I1206 14:09:33.833620 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e63f666-ace4-417b-ab05-50be328e8a64-bundle" (OuterVolumeSpecName: "bundle") pod "3e63f666-ace4-417b-ab05-50be328e8a64" (UID: "3e63f666-ace4-417b-ab05-50be328e8a64"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:09:33 crc kubenswrapper[4706]: I1206 14:09:33.836649 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e63f666-ace4-417b-ab05-50be328e8a64-kube-api-access-759k7" (OuterVolumeSpecName: "kube-api-access-759k7") pod "3e63f666-ace4-417b-ab05-50be328e8a64" (UID: "3e63f666-ace4-417b-ab05-50be328e8a64"). InnerVolumeSpecName "kube-api-access-759k7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:09:33 crc kubenswrapper[4706]: I1206 14:09:33.857148 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e63f666-ace4-417b-ab05-50be328e8a64-util" (OuterVolumeSpecName: "util") pod "3e63f666-ace4-417b-ab05-50be328e8a64" (UID: "3e63f666-ace4-417b-ab05-50be328e8a64"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:09:33 crc kubenswrapper[4706]: I1206 14:09:33.930238 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-759k7\" (UniqueName: \"kubernetes.io/projected/3e63f666-ace4-417b-ab05-50be328e8a64-kube-api-access-759k7\") on node \"crc\" DevicePath \"\"" Dec 06 14:09:33 crc kubenswrapper[4706]: I1206 14:09:33.930270 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3e63f666-ace4-417b-ab05-50be328e8a64-util\") on node \"crc\" DevicePath \"\"" Dec 06 14:09:33 crc kubenswrapper[4706]: I1206 14:09:33.930282 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3e63f666-ace4-417b-ab05-50be328e8a64-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 14:09:34 crc kubenswrapper[4706]: I1206 14:09:34.370841 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" event={"ID":"3e63f666-ace4-417b-ab05-50be328e8a64","Type":"ContainerDied","Data":"80c136375d81048ee2eb052060f18436929332bdb3b117ddc739a22d130d57d5"} Dec 06 14:09:34 crc kubenswrapper[4706]: I1206 14:09:34.370898 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80c136375d81048ee2eb052060f18436929332bdb3b117ddc739a22d130d57d5" Dec 06 14:09:34 crc kubenswrapper[4706]: I1206 14:09:34.370982 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.041844 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw"] Dec 06 14:09:38 crc kubenswrapper[4706]: E1206 14:09:38.042241 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e63f666-ace4-417b-ab05-50be328e8a64" containerName="util" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.042252 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e63f666-ace4-417b-ab05-50be328e8a64" containerName="util" Dec 06 14:09:38 crc kubenswrapper[4706]: E1206 14:09:38.042266 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e63f666-ace4-417b-ab05-50be328e8a64" containerName="pull" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.042272 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e63f666-ace4-417b-ab05-50be328e8a64" containerName="pull" Dec 06 14:09:38 crc kubenswrapper[4706]: E1206 14:09:38.042288 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e63f666-ace4-417b-ab05-50be328e8a64" containerName="extract" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.042294 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e63f666-ace4-417b-ab05-50be328e8a64" containerName="extract" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.042406 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e63f666-ace4-417b-ab05-50be328e8a64" containerName="extract" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.042777 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.044564 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-dhb7c" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.045012 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.045573 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.129531 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw"] Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.204292 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7081ae6-22a4-485d-93af-0662f73654f3-webhook-cert\") pod \"mariadb-operator-controller-manager-759c788948-76bcw\" (UID: \"a7081ae6-22a4-485d-93af-0662f73654f3\") " pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.204344 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnwl6\" (UniqueName: \"kubernetes.io/projected/a7081ae6-22a4-485d-93af-0662f73654f3-kube-api-access-qnwl6\") pod \"mariadb-operator-controller-manager-759c788948-76bcw\" (UID: \"a7081ae6-22a4-485d-93af-0662f73654f3\") " pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.204373 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7081ae6-22a4-485d-93af-0662f73654f3-apiservice-cert\") pod \"mariadb-operator-controller-manager-759c788948-76bcw\" (UID: \"a7081ae6-22a4-485d-93af-0662f73654f3\") " pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.305721 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7081ae6-22a4-485d-93af-0662f73654f3-webhook-cert\") pod \"mariadb-operator-controller-manager-759c788948-76bcw\" (UID: \"a7081ae6-22a4-485d-93af-0662f73654f3\") " pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.305770 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnwl6\" (UniqueName: \"kubernetes.io/projected/a7081ae6-22a4-485d-93af-0662f73654f3-kube-api-access-qnwl6\") pod \"mariadb-operator-controller-manager-759c788948-76bcw\" (UID: \"a7081ae6-22a4-485d-93af-0662f73654f3\") " pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.305794 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7081ae6-22a4-485d-93af-0662f73654f3-apiservice-cert\") pod \"mariadb-operator-controller-manager-759c788948-76bcw\" (UID: \"a7081ae6-22a4-485d-93af-0662f73654f3\") " pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.311204 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7081ae6-22a4-485d-93af-0662f73654f3-apiservice-cert\") pod \"mariadb-operator-controller-manager-759c788948-76bcw\" (UID: \"a7081ae6-22a4-485d-93af-0662f73654f3\") " pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.311393 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7081ae6-22a4-485d-93af-0662f73654f3-webhook-cert\") pod \"mariadb-operator-controller-manager-759c788948-76bcw\" (UID: \"a7081ae6-22a4-485d-93af-0662f73654f3\") " pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.323963 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnwl6\" (UniqueName: \"kubernetes.io/projected/a7081ae6-22a4-485d-93af-0662f73654f3-kube-api-access-qnwl6\") pod \"mariadb-operator-controller-manager-759c788948-76bcw\" (UID: \"a7081ae6-22a4-485d-93af-0662f73654f3\") " pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.357357 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" Dec 06 14:09:38 crc kubenswrapper[4706]: I1206 14:09:38.756730 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw"] Dec 06 14:09:38 crc kubenswrapper[4706]: W1206 14:09:38.762807 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7081ae6_22a4_485d_93af_0662f73654f3.slice/crio-187aa8aee4ab2b433c077849f82dfd6509bdd4d5ef605d5d48147562de2e9972 WatchSource:0}: Error finding container 187aa8aee4ab2b433c077849f82dfd6509bdd4d5ef605d5d48147562de2e9972: Status 404 returned error can't find the container with id 187aa8aee4ab2b433c077849f82dfd6509bdd4d5ef605d5d48147562de2e9972 Dec 06 14:09:39 crc kubenswrapper[4706]: I1206 14:09:39.401385 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" event={"ID":"a7081ae6-22a4-485d-93af-0662f73654f3","Type":"ContainerStarted","Data":"187aa8aee4ab2b433c077849f82dfd6509bdd4d5ef605d5d48147562de2e9972"} Dec 06 14:09:43 crc kubenswrapper[4706]: I1206 14:09:43.431164 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" event={"ID":"a7081ae6-22a4-485d-93af-0662f73654f3","Type":"ContainerStarted","Data":"fe6eb7cc2cff3d138effb0cba8430cac91c2ea7aaaacb6fa5f74476dd0fb810f"} Dec 06 14:09:43 crc kubenswrapper[4706]: I1206 14:09:43.431584 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" Dec 06 14:09:43 crc kubenswrapper[4706]: I1206 14:09:43.448869 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" podStartSLOduration=2.040072948 podStartE2EDuration="5.448849217s" podCreationTimestamp="2025-12-06 14:09:38 +0000 UTC" firstStartedPulling="2025-12-06 14:09:38.765562391 +0000 UTC m=+821.601355248" lastFinishedPulling="2025-12-06 14:09:42.17433866 +0000 UTC m=+825.010131517" observedRunningTime="2025-12-06 14:09:43.443549757 +0000 UTC m=+826.279342614" watchObservedRunningTime="2025-12-06 14:09:43.448849217 +0000 UTC m=+826.284642074" Dec 06 14:09:48 crc kubenswrapper[4706]: I1206 14:09:48.363579 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" Dec 06 14:09:51 crc kubenswrapper[4706]: I1206 14:09:51.809932 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-jdf6j"] Dec 06 14:09:51 crc kubenswrapper[4706]: I1206 14:09:51.811289 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-jdf6j" Dec 06 14:09:51 crc kubenswrapper[4706]: I1206 14:09:51.813482 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-r52fc" Dec 06 14:09:51 crc kubenswrapper[4706]: I1206 14:09:51.827258 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-jdf6j"] Dec 06 14:09:52 crc kubenswrapper[4706]: I1206 14:09:52.002460 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6pks\" (UniqueName: \"kubernetes.io/projected/044af6e2-e173-4163-af6b-ca10a4cd4937-kube-api-access-v6pks\") pod \"infra-operator-index-jdf6j\" (UID: \"044af6e2-e173-4163-af6b-ca10a4cd4937\") " pod="openstack-operators/infra-operator-index-jdf6j" Dec 06 14:09:52 crc kubenswrapper[4706]: I1206 14:09:52.103282 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6pks\" (UniqueName: \"kubernetes.io/projected/044af6e2-e173-4163-af6b-ca10a4cd4937-kube-api-access-v6pks\") pod \"infra-operator-index-jdf6j\" (UID: \"044af6e2-e173-4163-af6b-ca10a4cd4937\") " pod="openstack-operators/infra-operator-index-jdf6j" Dec 06 14:09:52 crc kubenswrapper[4706]: I1206 14:09:52.124425 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6pks\" (UniqueName: \"kubernetes.io/projected/044af6e2-e173-4163-af6b-ca10a4cd4937-kube-api-access-v6pks\") pod \"infra-operator-index-jdf6j\" (UID: \"044af6e2-e173-4163-af6b-ca10a4cd4937\") " pod="openstack-operators/infra-operator-index-jdf6j" Dec 06 14:09:52 crc kubenswrapper[4706]: I1206 14:09:52.199054 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-jdf6j" Dec 06 14:09:52 crc kubenswrapper[4706]: I1206 14:09:52.432857 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-jdf6j"] Dec 06 14:09:52 crc kubenswrapper[4706]: I1206 14:09:52.492133 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-jdf6j" event={"ID":"044af6e2-e173-4163-af6b-ca10a4cd4937","Type":"ContainerStarted","Data":"04b3cc5f74c822a559f2dcbc62ee4d55c37744a30d3d2de209c0c6e17944dafe"} Dec 06 14:09:56 crc kubenswrapper[4706]: I1206 14:09:56.203497 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-jdf6j"] Dec 06 14:09:56 crc kubenswrapper[4706]: I1206 14:09:56.806615 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-7vllp"] Dec 06 14:09:56 crc kubenswrapper[4706]: I1206 14:09:56.808249 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-7vllp" Dec 06 14:09:56 crc kubenswrapper[4706]: I1206 14:09:56.823643 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-7vllp"] Dec 06 14:09:56 crc kubenswrapper[4706]: I1206 14:09:56.971770 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b6sc\" (UniqueName: \"kubernetes.io/projected/e5625341-f864-4891-87e4-833fbd15d8da-kube-api-access-2b6sc\") pod \"infra-operator-index-7vllp\" (UID: \"e5625341-f864-4891-87e4-833fbd15d8da\") " pod="openstack-operators/infra-operator-index-7vllp" Dec 06 14:09:57 crc kubenswrapper[4706]: I1206 14:09:57.073581 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b6sc\" (UniqueName: \"kubernetes.io/projected/e5625341-f864-4891-87e4-833fbd15d8da-kube-api-access-2b6sc\") pod \"infra-operator-index-7vllp\" (UID: \"e5625341-f864-4891-87e4-833fbd15d8da\") " pod="openstack-operators/infra-operator-index-7vllp" Dec 06 14:09:57 crc kubenswrapper[4706]: I1206 14:09:57.094186 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b6sc\" (UniqueName: \"kubernetes.io/projected/e5625341-f864-4891-87e4-833fbd15d8da-kube-api-access-2b6sc\") pod \"infra-operator-index-7vllp\" (UID: \"e5625341-f864-4891-87e4-833fbd15d8da\") " pod="openstack-operators/infra-operator-index-7vllp" Dec 06 14:09:57 crc kubenswrapper[4706]: I1206 14:09:57.173197 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-7vllp" Dec 06 14:09:57 crc kubenswrapper[4706]: I1206 14:09:57.666778 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-7vllp"] Dec 06 14:09:57 crc kubenswrapper[4706]: W1206 14:09:57.782218 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5625341_f864_4891_87e4_833fbd15d8da.slice/crio-6b5347ee62e11dba51d6ff45845a48405c3fe7afd92a96bf451c561781b18a57 WatchSource:0}: Error finding container 6b5347ee62e11dba51d6ff45845a48405c3fe7afd92a96bf451c561781b18a57: Status 404 returned error can't find the container with id 6b5347ee62e11dba51d6ff45845a48405c3fe7afd92a96bf451c561781b18a57 Dec 06 14:09:58 crc kubenswrapper[4706]: I1206 14:09:58.526930 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-7vllp" event={"ID":"e5625341-f864-4891-87e4-833fbd15d8da","Type":"ContainerStarted","Data":"6b5347ee62e11dba51d6ff45845a48405c3fe7afd92a96bf451c561781b18a57"} Dec 06 14:09:58 crc kubenswrapper[4706]: I1206 14:09:58.534174 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-jdf6j" event={"ID":"044af6e2-e173-4163-af6b-ca10a4cd4937","Type":"ContainerStarted","Data":"60c5988faa3fab761e4f2b3be5a9c6d58e220b0756ea8a1f2e13f9b4a699b038"} Dec 06 14:09:58 crc kubenswrapper[4706]: I1206 14:09:58.534302 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-jdf6j" podUID="044af6e2-e173-4163-af6b-ca10a4cd4937" containerName="registry-server" containerID="cri-o://60c5988faa3fab761e4f2b3be5a9c6d58e220b0756ea8a1f2e13f9b4a699b038" gracePeriod=2 Dec 06 14:09:58 crc kubenswrapper[4706]: I1206 14:09:58.556739 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-jdf6j" podStartSLOduration=2.209718948 podStartE2EDuration="7.556718741s" podCreationTimestamp="2025-12-06 14:09:51 +0000 UTC" firstStartedPulling="2025-12-06 14:09:52.442427036 +0000 UTC m=+835.278219893" lastFinishedPulling="2025-12-06 14:09:57.789426819 +0000 UTC m=+840.625219686" observedRunningTime="2025-12-06 14:09:58.552972331 +0000 UTC m=+841.388765198" watchObservedRunningTime="2025-12-06 14:09:58.556718741 +0000 UTC m=+841.392511608" Dec 06 14:09:59 crc kubenswrapper[4706]: I1206 14:09:59.491783 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-jdf6j" Dec 06 14:09:59 crc kubenswrapper[4706]: I1206 14:09:59.503780 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6pks\" (UniqueName: \"kubernetes.io/projected/044af6e2-e173-4163-af6b-ca10a4cd4937-kube-api-access-v6pks\") pod \"044af6e2-e173-4163-af6b-ca10a4cd4937\" (UID: \"044af6e2-e173-4163-af6b-ca10a4cd4937\") " Dec 06 14:09:59 crc kubenswrapper[4706]: I1206 14:09:59.513670 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/044af6e2-e173-4163-af6b-ca10a4cd4937-kube-api-access-v6pks" (OuterVolumeSpecName: "kube-api-access-v6pks") pod "044af6e2-e173-4163-af6b-ca10a4cd4937" (UID: "044af6e2-e173-4163-af6b-ca10a4cd4937"). InnerVolumeSpecName "kube-api-access-v6pks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:09:59 crc kubenswrapper[4706]: I1206 14:09:59.546690 4706 generic.go:334] "Generic (PLEG): container finished" podID="044af6e2-e173-4163-af6b-ca10a4cd4937" containerID="60c5988faa3fab761e4f2b3be5a9c6d58e220b0756ea8a1f2e13f9b4a699b038" exitCode=0 Dec 06 14:09:59 crc kubenswrapper[4706]: I1206 14:09:59.546737 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-jdf6j" event={"ID":"044af6e2-e173-4163-af6b-ca10a4cd4937","Type":"ContainerDied","Data":"60c5988faa3fab761e4f2b3be5a9c6d58e220b0756ea8a1f2e13f9b4a699b038"} Dec 06 14:09:59 crc kubenswrapper[4706]: I1206 14:09:59.546797 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-jdf6j" event={"ID":"044af6e2-e173-4163-af6b-ca10a4cd4937","Type":"ContainerDied","Data":"04b3cc5f74c822a559f2dcbc62ee4d55c37744a30d3d2de209c0c6e17944dafe"} Dec 06 14:09:59 crc kubenswrapper[4706]: I1206 14:09:59.546796 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-jdf6j" Dec 06 14:09:59 crc kubenswrapper[4706]: I1206 14:09:59.546816 4706 scope.go:117] "RemoveContainer" containerID="60c5988faa3fab761e4f2b3be5a9c6d58e220b0756ea8a1f2e13f9b4a699b038" Dec 06 14:09:59 crc kubenswrapper[4706]: I1206 14:09:59.547845 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-7vllp" event={"ID":"e5625341-f864-4891-87e4-833fbd15d8da","Type":"ContainerStarted","Data":"dbec3957b7f16b9db17d46328986557b6b4b0e235d7b57e7dd9eddc7080a3bc1"} Dec 06 14:09:59 crc kubenswrapper[4706]: I1206 14:09:59.574360 4706 scope.go:117] "RemoveContainer" containerID="60c5988faa3fab761e4f2b3be5a9c6d58e220b0756ea8a1f2e13f9b4a699b038" Dec 06 14:09:59 crc kubenswrapper[4706]: E1206 14:09:59.575508 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60c5988faa3fab761e4f2b3be5a9c6d58e220b0756ea8a1f2e13f9b4a699b038\": container with ID starting with 60c5988faa3fab761e4f2b3be5a9c6d58e220b0756ea8a1f2e13f9b4a699b038 not found: ID does not exist" containerID="60c5988faa3fab761e4f2b3be5a9c6d58e220b0756ea8a1f2e13f9b4a699b038" Dec 06 14:09:59 crc kubenswrapper[4706]: I1206 14:09:59.575541 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60c5988faa3fab761e4f2b3be5a9c6d58e220b0756ea8a1f2e13f9b4a699b038"} err="failed to get container status \"60c5988faa3fab761e4f2b3be5a9c6d58e220b0756ea8a1f2e13f9b4a699b038\": rpc error: code = NotFound desc = could not find container \"60c5988faa3fab761e4f2b3be5a9c6d58e220b0756ea8a1f2e13f9b4a699b038\": container with ID starting with 60c5988faa3fab761e4f2b3be5a9c6d58e220b0756ea8a1f2e13f9b4a699b038 not found: ID does not exist" Dec 06 14:09:59 crc kubenswrapper[4706]: I1206 14:09:59.575948 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-7vllp" podStartSLOduration=2.5898733 podStartE2EDuration="3.575925951s" podCreationTimestamp="2025-12-06 14:09:56 +0000 UTC" firstStartedPulling="2025-12-06 14:09:57.787741394 +0000 UTC m=+840.623534271" lastFinishedPulling="2025-12-06 14:09:58.773794065 +0000 UTC m=+841.609586922" observedRunningTime="2025-12-06 14:09:59.566221524 +0000 UTC m=+842.402014411" watchObservedRunningTime="2025-12-06 14:09:59.575925951 +0000 UTC m=+842.411718818" Dec 06 14:09:59 crc kubenswrapper[4706]: I1206 14:09:59.586934 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-jdf6j"] Dec 06 14:09:59 crc kubenswrapper[4706]: I1206 14:09:59.592204 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-jdf6j"] Dec 06 14:09:59 crc kubenswrapper[4706]: I1206 14:09:59.610160 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6pks\" (UniqueName: \"kubernetes.io/projected/044af6e2-e173-4163-af6b-ca10a4cd4937-kube-api-access-v6pks\") on node \"crc\" DevicePath \"\"" Dec 06 14:10:01 crc kubenswrapper[4706]: I1206 14:10:01.522257 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="044af6e2-e173-4163-af6b-ca10a4cd4937" path="/var/lib/kubelet/pods/044af6e2-e173-4163-af6b-ca10a4cd4937/volumes" Dec 06 14:10:07 crc kubenswrapper[4706]: I1206 14:10:07.174153 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-7vllp" Dec 06 14:10:07 crc kubenswrapper[4706]: I1206 14:10:07.174740 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-7vllp" Dec 06 14:10:07 crc kubenswrapper[4706]: I1206 14:10:07.203645 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-7vllp" Dec 06 14:10:07 crc kubenswrapper[4706]: I1206 14:10:07.634213 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-7vllp" Dec 06 14:10:10 crc kubenswrapper[4706]: I1206 14:10:10.067481 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d"] Dec 06 14:10:10 crc kubenswrapper[4706]: E1206 14:10:10.068144 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044af6e2-e173-4163-af6b-ca10a4cd4937" containerName="registry-server" Dec 06 14:10:10 crc kubenswrapper[4706]: I1206 14:10:10.068166 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="044af6e2-e173-4163-af6b-ca10a4cd4937" containerName="registry-server" Dec 06 14:10:10 crc kubenswrapper[4706]: I1206 14:10:10.068579 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="044af6e2-e173-4163-af6b-ca10a4cd4937" containerName="registry-server" Dec 06 14:10:10 crc kubenswrapper[4706]: I1206 14:10:10.070261 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" Dec 06 14:10:10 crc kubenswrapper[4706]: I1206 14:10:10.073431 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-fc6v2" Dec 06 14:10:10 crc kubenswrapper[4706]: I1206 14:10:10.086921 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d"] Dec 06 14:10:10 crc kubenswrapper[4706]: I1206 14:10:10.147417 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t66pv\" (UniqueName: \"kubernetes.io/projected/44f63b09-00e7-45a5-b5c1-ffb282316231-kube-api-access-t66pv\") pod \"7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d\" (UID: \"44f63b09-00e7-45a5-b5c1-ffb282316231\") " pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" Dec 06 14:10:10 crc kubenswrapper[4706]: I1206 14:10:10.147507 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44f63b09-00e7-45a5-b5c1-ffb282316231-bundle\") pod \"7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d\" (UID: \"44f63b09-00e7-45a5-b5c1-ffb282316231\") " pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" Dec 06 14:10:10 crc kubenswrapper[4706]: I1206 14:10:10.147809 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44f63b09-00e7-45a5-b5c1-ffb282316231-util\") pod \"7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d\" (UID: \"44f63b09-00e7-45a5-b5c1-ffb282316231\") " pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" Dec 06 14:10:10 crc kubenswrapper[4706]: I1206 14:10:10.248966 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t66pv\" (UniqueName: \"kubernetes.io/projected/44f63b09-00e7-45a5-b5c1-ffb282316231-kube-api-access-t66pv\") pod \"7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d\" (UID: \"44f63b09-00e7-45a5-b5c1-ffb282316231\") " pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" Dec 06 14:10:10 crc kubenswrapper[4706]: I1206 14:10:10.249024 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44f63b09-00e7-45a5-b5c1-ffb282316231-bundle\") pod \"7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d\" (UID: \"44f63b09-00e7-45a5-b5c1-ffb282316231\") " pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" Dec 06 14:10:10 crc kubenswrapper[4706]: I1206 14:10:10.249099 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44f63b09-00e7-45a5-b5c1-ffb282316231-util\") pod \"7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d\" (UID: \"44f63b09-00e7-45a5-b5c1-ffb282316231\") " pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" Dec 06 14:10:10 crc kubenswrapper[4706]: I1206 14:10:10.249757 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44f63b09-00e7-45a5-b5c1-ffb282316231-util\") pod \"7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d\" (UID: \"44f63b09-00e7-45a5-b5c1-ffb282316231\") " pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" Dec 06 14:10:10 crc kubenswrapper[4706]: I1206 14:10:10.250135 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44f63b09-00e7-45a5-b5c1-ffb282316231-bundle\") pod \"7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d\" (UID: \"44f63b09-00e7-45a5-b5c1-ffb282316231\") " pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" Dec 06 14:10:10 crc kubenswrapper[4706]: I1206 14:10:10.288885 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t66pv\" (UniqueName: \"kubernetes.io/projected/44f63b09-00e7-45a5-b5c1-ffb282316231-kube-api-access-t66pv\") pod \"7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d\" (UID: \"44f63b09-00e7-45a5-b5c1-ffb282316231\") " pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" Dec 06 14:10:10 crc kubenswrapper[4706]: I1206 14:10:10.408962 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" Dec 06 14:10:10 crc kubenswrapper[4706]: I1206 14:10:10.665260 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d"] Dec 06 14:10:10 crc kubenswrapper[4706]: W1206 14:10:10.675402 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44f63b09_00e7_45a5_b5c1_ffb282316231.slice/crio-23675a66e4eef38849610d46b50d8be954d9b801f2d430afe2a145f92a5961cb WatchSource:0}: Error finding container 23675a66e4eef38849610d46b50d8be954d9b801f2d430afe2a145f92a5961cb: Status 404 returned error can't find the container with id 23675a66e4eef38849610d46b50d8be954d9b801f2d430afe2a145f92a5961cb Dec 06 14:10:11 crc kubenswrapper[4706]: I1206 14:10:11.627945 4706 generic.go:334] "Generic (PLEG): container finished" podID="44f63b09-00e7-45a5-b5c1-ffb282316231" containerID="3aa34f564128962b20c74ee2bbadd559a82b9b733956c2fee62daced7e00c835" exitCode=0 Dec 06 14:10:11 crc kubenswrapper[4706]: I1206 14:10:11.628059 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" event={"ID":"44f63b09-00e7-45a5-b5c1-ffb282316231","Type":"ContainerDied","Data":"3aa34f564128962b20c74ee2bbadd559a82b9b733956c2fee62daced7e00c835"} Dec 06 14:10:11 crc kubenswrapper[4706]: I1206 14:10:11.628337 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" event={"ID":"44f63b09-00e7-45a5-b5c1-ffb282316231","Type":"ContainerStarted","Data":"23675a66e4eef38849610d46b50d8be954d9b801f2d430afe2a145f92a5961cb"} Dec 06 14:10:12 crc kubenswrapper[4706]: I1206 14:10:12.636444 4706 generic.go:334] "Generic (PLEG): container finished" podID="44f63b09-00e7-45a5-b5c1-ffb282316231" containerID="b3ce0d041bb17a241d4703ada9ad4f49f6e0f6ac810b3f5d933da58a9e0236ec" exitCode=0 Dec 06 14:10:12 crc kubenswrapper[4706]: I1206 14:10:12.636510 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" event={"ID":"44f63b09-00e7-45a5-b5c1-ffb282316231","Type":"ContainerDied","Data":"b3ce0d041bb17a241d4703ada9ad4f49f6e0f6ac810b3f5d933da58a9e0236ec"} Dec 06 14:10:13 crc kubenswrapper[4706]: I1206 14:10:13.653705 4706 generic.go:334] "Generic (PLEG): container finished" podID="44f63b09-00e7-45a5-b5c1-ffb282316231" containerID="7b784979a0eb31f9254a036717a835a82031cd97863423e2a7353a3578656358" exitCode=0 Dec 06 14:10:13 crc kubenswrapper[4706]: I1206 14:10:13.653806 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" event={"ID":"44f63b09-00e7-45a5-b5c1-ffb282316231","Type":"ContainerDied","Data":"7b784979a0eb31f9254a036717a835a82031cd97863423e2a7353a3578656358"} Dec 06 14:10:15 crc kubenswrapper[4706]: I1206 14:10:15.011281 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" Dec 06 14:10:15 crc kubenswrapper[4706]: I1206 14:10:15.123466 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44f63b09-00e7-45a5-b5c1-ffb282316231-util\") pod \"44f63b09-00e7-45a5-b5c1-ffb282316231\" (UID: \"44f63b09-00e7-45a5-b5c1-ffb282316231\") " Dec 06 14:10:15 crc kubenswrapper[4706]: I1206 14:10:15.123544 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t66pv\" (UniqueName: \"kubernetes.io/projected/44f63b09-00e7-45a5-b5c1-ffb282316231-kube-api-access-t66pv\") pod \"44f63b09-00e7-45a5-b5c1-ffb282316231\" (UID: \"44f63b09-00e7-45a5-b5c1-ffb282316231\") " Dec 06 14:10:15 crc kubenswrapper[4706]: I1206 14:10:15.123670 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44f63b09-00e7-45a5-b5c1-ffb282316231-bundle\") pod \"44f63b09-00e7-45a5-b5c1-ffb282316231\" (UID: \"44f63b09-00e7-45a5-b5c1-ffb282316231\") " Dec 06 14:10:15 crc kubenswrapper[4706]: I1206 14:10:15.126035 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44f63b09-00e7-45a5-b5c1-ffb282316231-bundle" (OuterVolumeSpecName: "bundle") pod "44f63b09-00e7-45a5-b5c1-ffb282316231" (UID: "44f63b09-00e7-45a5-b5c1-ffb282316231"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:10:15 crc kubenswrapper[4706]: I1206 14:10:15.129776 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44f63b09-00e7-45a5-b5c1-ffb282316231-kube-api-access-t66pv" (OuterVolumeSpecName: "kube-api-access-t66pv") pod "44f63b09-00e7-45a5-b5c1-ffb282316231" (UID: "44f63b09-00e7-45a5-b5c1-ffb282316231"). InnerVolumeSpecName "kube-api-access-t66pv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:10:15 crc kubenswrapper[4706]: I1206 14:10:15.160467 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44f63b09-00e7-45a5-b5c1-ffb282316231-util" (OuterVolumeSpecName: "util") pod "44f63b09-00e7-45a5-b5c1-ffb282316231" (UID: "44f63b09-00e7-45a5-b5c1-ffb282316231"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:10:15 crc kubenswrapper[4706]: I1206 14:10:15.225730 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44f63b09-00e7-45a5-b5c1-ffb282316231-util\") on node \"crc\" DevicePath \"\"" Dec 06 14:10:15 crc kubenswrapper[4706]: I1206 14:10:15.225782 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t66pv\" (UniqueName: \"kubernetes.io/projected/44f63b09-00e7-45a5-b5c1-ffb282316231-kube-api-access-t66pv\") on node \"crc\" DevicePath \"\"" Dec 06 14:10:15 crc kubenswrapper[4706]: I1206 14:10:15.225803 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44f63b09-00e7-45a5-b5c1-ffb282316231-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 14:10:15 crc kubenswrapper[4706]: I1206 14:10:15.677216 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" event={"ID":"44f63b09-00e7-45a5-b5c1-ffb282316231","Type":"ContainerDied","Data":"23675a66e4eef38849610d46b50d8be954d9b801f2d430afe2a145f92a5961cb"} Dec 06 14:10:15 crc kubenswrapper[4706]: I1206 14:10:15.677649 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23675a66e4eef38849610d46b50d8be954d9b801f2d430afe2a145f92a5961cb" Dec 06 14:10:15 crc kubenswrapper[4706]: I1206 14:10:15.677303 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.166949 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 06 14:10:24 crc kubenswrapper[4706]: E1206 14:10:24.167957 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44f63b09-00e7-45a5-b5c1-ffb282316231" containerName="extract" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.167980 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="44f63b09-00e7-45a5-b5c1-ffb282316231" containerName="extract" Dec 06 14:10:24 crc kubenswrapper[4706]: E1206 14:10:24.168013 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44f63b09-00e7-45a5-b5c1-ffb282316231" containerName="pull" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.168026 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="44f63b09-00e7-45a5-b5c1-ffb282316231" containerName="pull" Dec 06 14:10:24 crc kubenswrapper[4706]: E1206 14:10:24.168045 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44f63b09-00e7-45a5-b5c1-ffb282316231" containerName="util" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.168058 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="44f63b09-00e7-45a5-b5c1-ffb282316231" containerName="util" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.168232 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="44f63b09-00e7-45a5-b5c1-ffb282316231" containerName="extract" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.169234 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.171435 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"openshift-service-ca.crt" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.173433 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"openstack-config-data" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.173875 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"kube-root-ca.crt" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.174168 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"galera-openstack-dockercfg-4j5n7" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.174563 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"openstack-scripts" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.183153 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.197459 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.198680 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.202283 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.203365 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.236331 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.243090 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.263925 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.264019 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-config-data-default\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.264053 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.264149 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-operator-scripts\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.264208 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bhcc\" (UniqueName: \"kubernetes.io/projected/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-kube-api-access-9bhcc\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.264238 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-kolla-config\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.264342 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-config-data-default\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.264400 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.264434 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/41b63157-bdec-47c2-a4b0-228f89541a2a-config-data-generated\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.264457 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbgw2\" (UniqueName: \"kubernetes.io/projected/41b63157-bdec-47c2-a4b0-228f89541a2a-kube-api-access-rbgw2\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.264602 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-kolla-config\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.264670 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.365789 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.365865 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/41b63157-bdec-47c2-a4b0-228f89541a2a-config-data-generated\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.365932 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbgw2\" (UniqueName: \"kubernetes.io/projected/41b63157-bdec-47c2-a4b0-228f89541a2a-kube-api-access-rbgw2\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.366617 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/41b63157-bdec-47c2-a4b0-228f89541a2a-config-data-generated\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.366729 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/252d778a-eb1a-4519-9325-e79910544f1d-config-data-generated\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.366842 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-config-data-default\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.366946 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dl96\" (UniqueName: \"kubernetes.io/projected/252d778a-eb1a-4519-9325-e79910544f1d-kube-api-access-9dl96\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.367073 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-kolla-config\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.367134 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-kolla-config\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.367266 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.367414 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.367654 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") device mount path \"/mnt/openstack/pv07\"" pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.367688 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") device mount path \"/mnt/openstack/pv03\"" pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.367731 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-kolla-config\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.368236 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.376427 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-config-data-default\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.376595 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-operator-scripts\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.376639 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.376691 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-operator-scripts\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.376758 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.376787 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bhcc\" (UniqueName: \"kubernetes.io/projected/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-kube-api-access-9bhcc\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.376815 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-kolla-config\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.376851 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-config-data-default\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.376973 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.377559 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-config-data-default\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.377630 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-config-data-default\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.377671 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-kolla-config\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.378579 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-operator-scripts\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.383302 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbgw2\" (UniqueName: \"kubernetes.io/projected/41b63157-bdec-47c2-a4b0-228f89541a2a-kube-api-access-rbgw2\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.387441 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-2\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.389714 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.413114 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bhcc\" (UniqueName: \"kubernetes.io/projected/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-kube-api-access-9bhcc\") pod \"openstack-galera-0\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.478216 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-operator-scripts\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.478282 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.478356 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-config-data-default\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.478379 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/252d778a-eb1a-4519-9325-e79910544f1d-config-data-generated\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.478405 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dl96\" (UniqueName: \"kubernetes.io/projected/252d778a-eb1a-4519-9325-e79910544f1d-kube-api-access-9dl96\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.478473 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-kolla-config\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.478593 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") device mount path \"/mnt/openstack/pv09\"" pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.479376 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-kolla-config\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.479800 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-operator-scripts\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.479809 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/252d778a-eb1a-4519-9325-e79910544f1d-config-data-generated\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.480100 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-config-data-default\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.489716 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.502993 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.517055 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dl96\" (UniqueName: \"kubernetes.io/projected/252d778a-eb1a-4519-9325-e79910544f1d-kube-api-access-9dl96\") pod \"openstack-galera-1\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.522901 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.534381 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.718019 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s"] Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.718712 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.722414 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-hrjxs" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.722629 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.734018 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s"] Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.789253 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-apiservice-cert\") pod \"infra-operator-controller-manager-d96b7b8b9-qpq4s\" (UID: \"e430699f-156a-4c0c-b038-06a27b691852\") " pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.789355 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-webhook-cert\") pod \"infra-operator-controller-manager-d96b7b8b9-qpq4s\" (UID: \"e430699f-156a-4c0c-b038-06a27b691852\") " pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.789391 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zrn4\" (UniqueName: \"kubernetes.io/projected/e430699f-156a-4c0c-b038-06a27b691852-kube-api-access-7zrn4\") pod \"infra-operator-controller-manager-d96b7b8b9-qpq4s\" (UID: \"e430699f-156a-4c0c-b038-06a27b691852\") " pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.890659 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-apiservice-cert\") pod \"infra-operator-controller-manager-d96b7b8b9-qpq4s\" (UID: \"e430699f-156a-4c0c-b038-06a27b691852\") " pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.890730 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-webhook-cert\") pod \"infra-operator-controller-manager-d96b7b8b9-qpq4s\" (UID: \"e430699f-156a-4c0c-b038-06a27b691852\") " pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.890765 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zrn4\" (UniqueName: \"kubernetes.io/projected/e430699f-156a-4c0c-b038-06a27b691852-kube-api-access-7zrn4\") pod \"infra-operator-controller-manager-d96b7b8b9-qpq4s\" (UID: \"e430699f-156a-4c0c-b038-06a27b691852\") " pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.894768 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-apiservice-cert\") pod \"infra-operator-controller-manager-d96b7b8b9-qpq4s\" (UID: \"e430699f-156a-4c0c-b038-06a27b691852\") " pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.895780 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-webhook-cert\") pod \"infra-operator-controller-manager-d96b7b8b9-qpq4s\" (UID: \"e430699f-156a-4c0c-b038-06a27b691852\") " pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.907018 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zrn4\" (UniqueName: \"kubernetes.io/projected/e430699f-156a-4c0c-b038-06a27b691852-kube-api-access-7zrn4\") pod \"infra-operator-controller-manager-d96b7b8b9-qpq4s\" (UID: \"e430699f-156a-4c0c-b038-06a27b691852\") " pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" Dec 06 14:10:24 crc kubenswrapper[4706]: I1206 14:10:24.992649 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 06 14:10:25 crc kubenswrapper[4706]: I1206 14:10:25.044628 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" Dec 06 14:10:25 crc kubenswrapper[4706]: I1206 14:10:25.072707 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 06 14:10:25 crc kubenswrapper[4706]: I1206 14:10:25.080917 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 06 14:10:25 crc kubenswrapper[4706]: I1206 14:10:25.333847 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s"] Dec 06 14:10:25 crc kubenswrapper[4706]: W1206 14:10:25.346176 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode430699f_156a_4c0c_b038_06a27b691852.slice/crio-8bbfdabd7738dc8cba307bf56a29ce24295c719e7979a26b4f7b395b237638ae WatchSource:0}: Error finding container 8bbfdabd7738dc8cba307bf56a29ce24295c719e7979a26b4f7b395b237638ae: Status 404 returned error can't find the container with id 8bbfdabd7738dc8cba307bf56a29ce24295c719e7979a26b4f7b395b237638ae Dec 06 14:10:25 crc kubenswrapper[4706]: I1206 14:10:25.745619 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" event={"ID":"e430699f-156a-4c0c-b038-06a27b691852","Type":"ContainerStarted","Data":"8bbfdabd7738dc8cba307bf56a29ce24295c719e7979a26b4f7b395b237638ae"} Dec 06 14:10:25 crc kubenswrapper[4706]: I1206 14:10:25.747293 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"252d778a-eb1a-4519-9325-e79910544f1d","Type":"ContainerStarted","Data":"16684f78471591fa81eefb61f91abecf5e00c05be216dd56984380415a77c37f"} Dec 06 14:10:25 crc kubenswrapper[4706]: I1206 14:10:25.749038 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"a86ee400-6a8a-4cc6-bdea-62a40954a8b2","Type":"ContainerStarted","Data":"4db16c5d9b40a1c1ab2ba870e2692f4c045b4ddc9325d308503a5ee334dcf9f5"} Dec 06 14:10:25 crc kubenswrapper[4706]: I1206 14:10:25.752175 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"41b63157-bdec-47c2-a4b0-228f89541a2a","Type":"ContainerStarted","Data":"c4c34b6ec3c3851b309aac8ca34f4844b412945e3c33c811a238d93cf71a30e0"} Dec 06 14:10:34 crc kubenswrapper[4706]: I1206 14:10:34.820859 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" event={"ID":"e430699f-156a-4c0c-b038-06a27b691852","Type":"ContainerStarted","Data":"1a4db1c512a94198a3f6a86efb734f507e64d2c61348277262b04f52e017216e"} Dec 06 14:10:34 crc kubenswrapper[4706]: I1206 14:10:34.821459 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" Dec 06 14:10:34 crc kubenswrapper[4706]: I1206 14:10:34.823517 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"252d778a-eb1a-4519-9325-e79910544f1d","Type":"ContainerStarted","Data":"5041e6e52899dbad5204aedfe6bd40b7b14f8fa0b7a3bf46c4500dcc5b75f5df"} Dec 06 14:10:34 crc kubenswrapper[4706]: I1206 14:10:34.825243 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"a86ee400-6a8a-4cc6-bdea-62a40954a8b2","Type":"ContainerStarted","Data":"a591760eadcadf8809da63c6dd9d5afaaf5320cb90aa24fcd946f35cb81242ae"} Dec 06 14:10:34 crc kubenswrapper[4706]: I1206 14:10:34.826766 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"41b63157-bdec-47c2-a4b0-228f89541a2a","Type":"ContainerStarted","Data":"9553942fd0f85b5353637d9514ac2d86b968c0a0a74f85f4a90194f943c222fb"} Dec 06 14:10:34 crc kubenswrapper[4706]: I1206 14:10:34.846352 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" podStartSLOduration=2.142200483 podStartE2EDuration="10.846333262s" podCreationTimestamp="2025-12-06 14:10:24 +0000 UTC" firstStartedPulling="2025-12-06 14:10:25.354156445 +0000 UTC m=+868.189949342" lastFinishedPulling="2025-12-06 14:10:34.058289254 +0000 UTC m=+876.894082121" observedRunningTime="2025-12-06 14:10:34.83983864 +0000 UTC m=+877.675631497" watchObservedRunningTime="2025-12-06 14:10:34.846333262 +0000 UTC m=+877.682126119" Dec 06 14:10:41 crc kubenswrapper[4706]: I1206 14:10:41.879351 4706 generic.go:334] "Generic (PLEG): container finished" podID="a86ee400-6a8a-4cc6-bdea-62a40954a8b2" containerID="a591760eadcadf8809da63c6dd9d5afaaf5320cb90aa24fcd946f35cb81242ae" exitCode=0 Dec 06 14:10:41 crc kubenswrapper[4706]: I1206 14:10:41.879453 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"a86ee400-6a8a-4cc6-bdea-62a40954a8b2","Type":"ContainerDied","Data":"a591760eadcadf8809da63c6dd9d5afaaf5320cb90aa24fcd946f35cb81242ae"} Dec 06 14:10:41 crc kubenswrapper[4706]: I1206 14:10:41.883042 4706 generic.go:334] "Generic (PLEG): container finished" podID="41b63157-bdec-47c2-a4b0-228f89541a2a" containerID="9553942fd0f85b5353637d9514ac2d86b968c0a0a74f85f4a90194f943c222fb" exitCode=0 Dec 06 14:10:41 crc kubenswrapper[4706]: I1206 14:10:41.883177 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"41b63157-bdec-47c2-a4b0-228f89541a2a","Type":"ContainerDied","Data":"9553942fd0f85b5353637d9514ac2d86b968c0a0a74f85f4a90194f943c222fb"} Dec 06 14:10:41 crc kubenswrapper[4706]: I1206 14:10:41.885930 4706 generic.go:334] "Generic (PLEG): container finished" podID="252d778a-eb1a-4519-9325-e79910544f1d" containerID="5041e6e52899dbad5204aedfe6bd40b7b14f8fa0b7a3bf46c4500dcc5b75f5df" exitCode=0 Dec 06 14:10:41 crc kubenswrapper[4706]: I1206 14:10:41.885979 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"252d778a-eb1a-4519-9325-e79910544f1d","Type":"ContainerDied","Data":"5041e6e52899dbad5204aedfe6bd40b7b14f8fa0b7a3bf46c4500dcc5b75f5df"} Dec 06 14:10:42 crc kubenswrapper[4706]: I1206 14:10:42.895611 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"a86ee400-6a8a-4cc6-bdea-62a40954a8b2","Type":"ContainerStarted","Data":"fafab63b20f39590bf945c4153224cf7d8639a7ffad58deffc95be89683694e0"} Dec 06 14:10:42 crc kubenswrapper[4706]: I1206 14:10:42.898183 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"41b63157-bdec-47c2-a4b0-228f89541a2a","Type":"ContainerStarted","Data":"56494e751042f56606058af3e4fca29a89a08904e452133514ddd5ed3269298a"} Dec 06 14:10:42 crc kubenswrapper[4706]: I1206 14:10:42.901789 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"252d778a-eb1a-4519-9325-e79910544f1d","Type":"ContainerStarted","Data":"a92d440a79be04659f5deb197d5547b6bdfbb3797c5ad7c4659c397e557c4cf9"} Dec 06 14:10:42 crc kubenswrapper[4706]: I1206 14:10:42.918729 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/openstack-galera-0" podStartSLOduration=10.762643482 podStartE2EDuration="19.918711505s" podCreationTimestamp="2025-12-06 14:10:23 +0000 UTC" firstStartedPulling="2025-12-06 14:10:25.001397269 +0000 UTC m=+867.837190156" lastFinishedPulling="2025-12-06 14:10:34.157465312 +0000 UTC m=+876.993258179" observedRunningTime="2025-12-06 14:10:42.915192382 +0000 UTC m=+885.750985249" watchObservedRunningTime="2025-12-06 14:10:42.918711505 +0000 UTC m=+885.754504372" Dec 06 14:10:42 crc kubenswrapper[4706]: I1206 14:10:42.939763 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/openstack-galera-2" podStartSLOduration=10.862835057 podStartE2EDuration="19.939750493s" podCreationTimestamp="2025-12-06 14:10:23 +0000 UTC" firstStartedPulling="2025-12-06 14:10:25.077212787 +0000 UTC m=+867.913005654" lastFinishedPulling="2025-12-06 14:10:34.154128233 +0000 UTC m=+876.989921090" observedRunningTime="2025-12-06 14:10:42.936793994 +0000 UTC m=+885.772586901" watchObservedRunningTime="2025-12-06 14:10:42.939750493 +0000 UTC m=+885.775543360" Dec 06 14:10:42 crc kubenswrapper[4706]: I1206 14:10:42.964423 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/openstack-galera-1" podStartSLOduration=10.831392244 podStartE2EDuration="19.964403216s" podCreationTimestamp="2025-12-06 14:10:23 +0000 UTC" firstStartedPulling="2025-12-06 14:10:25.07767929 +0000 UTC m=+867.913472147" lastFinishedPulling="2025-12-06 14:10:34.210690262 +0000 UTC m=+877.046483119" observedRunningTime="2025-12-06 14:10:42.960447511 +0000 UTC m=+885.796240368" watchObservedRunningTime="2025-12-06 14:10:42.964403216 +0000 UTC m=+885.800196083" Dec 06 14:10:44 crc kubenswrapper[4706]: I1206 14:10:44.489953 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:44 crc kubenswrapper[4706]: I1206 14:10:44.490022 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:10:44 crc kubenswrapper[4706]: I1206 14:10:44.524016 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:44 crc kubenswrapper[4706]: I1206 14:10:44.524166 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:44 crc kubenswrapper[4706]: I1206 14:10:44.536031 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:44 crc kubenswrapper[4706]: I1206 14:10:44.536089 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:10:45 crc kubenswrapper[4706]: I1206 14:10:45.050177 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" Dec 06 14:10:45 crc kubenswrapper[4706]: I1206 14:10:45.717060 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 06 14:10:45 crc kubenswrapper[4706]: I1206 14:10:45.717884 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Dec 06 14:10:45 crc kubenswrapper[4706]: I1206 14:10:45.720060 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"memcached-memcached-dockercfg-kxlgh" Dec 06 14:10:45 crc kubenswrapper[4706]: I1206 14:10:45.720146 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"memcached-config-data" Dec 06 14:10:45 crc kubenswrapper[4706]: I1206 14:10:45.731153 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 06 14:10:45 crc kubenswrapper[4706]: I1206 14:10:45.823955 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/76a993fd-482d-4efd-938a-2014dda4aca1-kolla-config\") pod \"memcached-0\" (UID: \"76a993fd-482d-4efd-938a-2014dda4aca1\") " pod="cinder-kuttl-tests/memcached-0" Dec 06 14:10:45 crc kubenswrapper[4706]: I1206 14:10:45.824033 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76a993fd-482d-4efd-938a-2014dda4aca1-config-data\") pod \"memcached-0\" (UID: \"76a993fd-482d-4efd-938a-2014dda4aca1\") " pod="cinder-kuttl-tests/memcached-0" Dec 06 14:10:45 crc kubenswrapper[4706]: I1206 14:10:45.824095 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vvqb\" (UniqueName: \"kubernetes.io/projected/76a993fd-482d-4efd-938a-2014dda4aca1-kube-api-access-5vvqb\") pod \"memcached-0\" (UID: \"76a993fd-482d-4efd-938a-2014dda4aca1\") " pod="cinder-kuttl-tests/memcached-0" Dec 06 14:10:45 crc kubenswrapper[4706]: I1206 14:10:45.925033 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76a993fd-482d-4efd-938a-2014dda4aca1-config-data\") pod \"memcached-0\" (UID: \"76a993fd-482d-4efd-938a-2014dda4aca1\") " pod="cinder-kuttl-tests/memcached-0" Dec 06 14:10:45 crc kubenswrapper[4706]: I1206 14:10:45.925993 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76a993fd-482d-4efd-938a-2014dda4aca1-config-data\") pod \"memcached-0\" (UID: \"76a993fd-482d-4efd-938a-2014dda4aca1\") " pod="cinder-kuttl-tests/memcached-0" Dec 06 14:10:45 crc kubenswrapper[4706]: I1206 14:10:45.927224 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vvqb\" (UniqueName: \"kubernetes.io/projected/76a993fd-482d-4efd-938a-2014dda4aca1-kube-api-access-5vvqb\") pod \"memcached-0\" (UID: \"76a993fd-482d-4efd-938a-2014dda4aca1\") " pod="cinder-kuttl-tests/memcached-0" Dec 06 14:10:45 crc kubenswrapper[4706]: I1206 14:10:45.927337 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/76a993fd-482d-4efd-938a-2014dda4aca1-kolla-config\") pod \"memcached-0\" (UID: \"76a993fd-482d-4efd-938a-2014dda4aca1\") " pod="cinder-kuttl-tests/memcached-0" Dec 06 14:10:45 crc kubenswrapper[4706]: I1206 14:10:45.928046 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/76a993fd-482d-4efd-938a-2014dda4aca1-kolla-config\") pod \"memcached-0\" (UID: \"76a993fd-482d-4efd-938a-2014dda4aca1\") " pod="cinder-kuttl-tests/memcached-0" Dec 06 14:10:45 crc kubenswrapper[4706]: I1206 14:10:45.951145 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vvqb\" (UniqueName: \"kubernetes.io/projected/76a993fd-482d-4efd-938a-2014dda4aca1-kube-api-access-5vvqb\") pod \"memcached-0\" (UID: \"76a993fd-482d-4efd-938a-2014dda4aca1\") " pod="cinder-kuttl-tests/memcached-0" Dec 06 14:10:46 crc kubenswrapper[4706]: I1206 14:10:46.035906 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Dec 06 14:10:46 crc kubenswrapper[4706]: I1206 14:10:46.494453 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 06 14:10:46 crc kubenswrapper[4706]: W1206 14:10:46.498105 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76a993fd_482d_4efd_938a_2014dda4aca1.slice/crio-0c6aef584ebedf1f371a4af402cc6896b08e5195e233691a8faf4830ef5705dc WatchSource:0}: Error finding container 0c6aef584ebedf1f371a4af402cc6896b08e5195e233691a8faf4830ef5705dc: Status 404 returned error can't find the container with id 0c6aef584ebedf1f371a4af402cc6896b08e5195e233691a8faf4830ef5705dc Dec 06 14:10:46 crc kubenswrapper[4706]: I1206 14:10:46.658768 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-m4592"] Dec 06 14:10:46 crc kubenswrapper[4706]: I1206 14:10:46.659984 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-m4592" Dec 06 14:10:46 crc kubenswrapper[4706]: I1206 14:10:46.663178 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-zbzwx" Dec 06 14:10:46 crc kubenswrapper[4706]: I1206 14:10:46.665973 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-m4592"] Dec 06 14:10:46 crc kubenswrapper[4706]: I1206 14:10:46.839172 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nmph\" (UniqueName: \"kubernetes.io/projected/9160e5ea-636c-4924-8ba1-1dfd13dfa0cc-kube-api-access-7nmph\") pod \"rabbitmq-cluster-operator-index-m4592\" (UID: \"9160e5ea-636c-4924-8ba1-1dfd13dfa0cc\") " pod="openstack-operators/rabbitmq-cluster-operator-index-m4592" Dec 06 14:10:46 crc kubenswrapper[4706]: I1206 14:10:46.932092 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"76a993fd-482d-4efd-938a-2014dda4aca1","Type":"ContainerStarted","Data":"0c6aef584ebedf1f371a4af402cc6896b08e5195e233691a8faf4830ef5705dc"} Dec 06 14:10:46 crc kubenswrapper[4706]: I1206 14:10:46.940072 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nmph\" (UniqueName: \"kubernetes.io/projected/9160e5ea-636c-4924-8ba1-1dfd13dfa0cc-kube-api-access-7nmph\") pod \"rabbitmq-cluster-operator-index-m4592\" (UID: \"9160e5ea-636c-4924-8ba1-1dfd13dfa0cc\") " pod="openstack-operators/rabbitmq-cluster-operator-index-m4592" Dec 06 14:10:46 crc kubenswrapper[4706]: I1206 14:10:46.977710 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nmph\" (UniqueName: \"kubernetes.io/projected/9160e5ea-636c-4924-8ba1-1dfd13dfa0cc-kube-api-access-7nmph\") pod \"rabbitmq-cluster-operator-index-m4592\" (UID: \"9160e5ea-636c-4924-8ba1-1dfd13dfa0cc\") " pod="openstack-operators/rabbitmq-cluster-operator-index-m4592" Dec 06 14:10:47 crc kubenswrapper[4706]: I1206 14:10:47.273205 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-m4592" Dec 06 14:10:47 crc kubenswrapper[4706]: I1206 14:10:47.475779 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-m4592"] Dec 06 14:10:47 crc kubenswrapper[4706]: I1206 14:10:47.940451 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-m4592" event={"ID":"9160e5ea-636c-4924-8ba1-1dfd13dfa0cc","Type":"ContainerStarted","Data":"054abdabb0c0f696fc5c971b9b1acf846fafaefedaf776834798bb36fa31f47d"} Dec 06 14:10:48 crc kubenswrapper[4706]: I1206 14:10:48.948863 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"76a993fd-482d-4efd-938a-2014dda4aca1","Type":"ContainerStarted","Data":"9f53871a8284673c6d9414d0374c0bf6afe88dd86fc2f4264ab391d6f28557f2"} Dec 06 14:10:48 crc kubenswrapper[4706]: I1206 14:10:48.949249 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/memcached-0" Dec 06 14:10:48 crc kubenswrapper[4706]: I1206 14:10:48.969249 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/memcached-0" podStartSLOduration=2.123481773 podStartE2EDuration="3.969225182s" podCreationTimestamp="2025-12-06 14:10:45 +0000 UTC" firstStartedPulling="2025-12-06 14:10:46.500280963 +0000 UTC m=+889.336073820" lastFinishedPulling="2025-12-06 14:10:48.346024372 +0000 UTC m=+891.181817229" observedRunningTime="2025-12-06 14:10:48.962745951 +0000 UTC m=+891.798538808" watchObservedRunningTime="2025-12-06 14:10:48.969225182 +0000 UTC m=+891.805018039" Dec 06 14:10:50 crc kubenswrapper[4706]: I1206 14:10:50.655410 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:50 crc kubenswrapper[4706]: I1206 14:10:50.754104 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:10:50 crc kubenswrapper[4706]: I1206 14:10:50.854648 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-m4592"] Dec 06 14:10:50 crc kubenswrapper[4706]: E1206 14:10:50.965369 4706 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.83:36768->38.102.83.83:34293: write tcp 38.102.83.83:36768->38.102.83.83:34293: write: broken pipe Dec 06 14:10:51 crc kubenswrapper[4706]: I1206 14:10:51.473255 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-6xsvs"] Dec 06 14:10:51 crc kubenswrapper[4706]: I1206 14:10:51.473991 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-6xsvs" Dec 06 14:10:51 crc kubenswrapper[4706]: I1206 14:10:51.485748 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-6xsvs"] Dec 06 14:10:51 crc kubenswrapper[4706]: I1206 14:10:51.607092 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2lj6\" (UniqueName: \"kubernetes.io/projected/352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9-kube-api-access-j2lj6\") pod \"rabbitmq-cluster-operator-index-6xsvs\" (UID: \"352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9\") " pod="openstack-operators/rabbitmq-cluster-operator-index-6xsvs" Dec 06 14:10:51 crc kubenswrapper[4706]: I1206 14:10:51.708769 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2lj6\" (UniqueName: \"kubernetes.io/projected/352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9-kube-api-access-j2lj6\") pod \"rabbitmq-cluster-operator-index-6xsvs\" (UID: \"352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9\") " pod="openstack-operators/rabbitmq-cluster-operator-index-6xsvs" Dec 06 14:10:51 crc kubenswrapper[4706]: I1206 14:10:51.725521 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2lj6\" (UniqueName: \"kubernetes.io/projected/352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9-kube-api-access-j2lj6\") pod \"rabbitmq-cluster-operator-index-6xsvs\" (UID: \"352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9\") " pod="openstack-operators/rabbitmq-cluster-operator-index-6xsvs" Dec 06 14:10:51 crc kubenswrapper[4706]: I1206 14:10:51.789530 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-6xsvs" Dec 06 14:10:51 crc kubenswrapper[4706]: I1206 14:10:51.966406 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-m4592" event={"ID":"9160e5ea-636c-4924-8ba1-1dfd13dfa0cc","Type":"ContainerStarted","Data":"7f1b924815e7d644d23cbb802884fe1a940f8ecc413e11e0ee7a3f6ddf434af7"} Dec 06 14:10:51 crc kubenswrapper[4706]: I1206 14:10:51.966964 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-m4592" podUID="9160e5ea-636c-4924-8ba1-1dfd13dfa0cc" containerName="registry-server" containerID="cri-o://7f1b924815e7d644d23cbb802884fe1a940f8ecc413e11e0ee7a3f6ddf434af7" gracePeriod=2 Dec 06 14:10:51 crc kubenswrapper[4706]: I1206 14:10:51.985810 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-m4592" podStartSLOduration=1.819322487 podStartE2EDuration="5.985794871s" podCreationTimestamp="2025-12-06 14:10:46 +0000 UTC" firstStartedPulling="2025-12-06 14:10:47.48764501 +0000 UTC m=+890.323437867" lastFinishedPulling="2025-12-06 14:10:51.654117394 +0000 UTC m=+894.489910251" observedRunningTime="2025-12-06 14:10:51.983759027 +0000 UTC m=+894.819551884" watchObservedRunningTime="2025-12-06 14:10:51.985794871 +0000 UTC m=+894.821587728" Dec 06 14:10:52 crc kubenswrapper[4706]: I1206 14:10:52.230550 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-6xsvs"] Dec 06 14:10:52 crc kubenswrapper[4706]: I1206 14:10:52.295350 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-m4592" Dec 06 14:10:52 crc kubenswrapper[4706]: I1206 14:10:52.416378 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nmph\" (UniqueName: \"kubernetes.io/projected/9160e5ea-636c-4924-8ba1-1dfd13dfa0cc-kube-api-access-7nmph\") pod \"9160e5ea-636c-4924-8ba1-1dfd13dfa0cc\" (UID: \"9160e5ea-636c-4924-8ba1-1dfd13dfa0cc\") " Dec 06 14:10:52 crc kubenswrapper[4706]: I1206 14:10:52.420977 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9160e5ea-636c-4924-8ba1-1dfd13dfa0cc-kube-api-access-7nmph" (OuterVolumeSpecName: "kube-api-access-7nmph") pod "9160e5ea-636c-4924-8ba1-1dfd13dfa0cc" (UID: "9160e5ea-636c-4924-8ba1-1dfd13dfa0cc"). InnerVolumeSpecName "kube-api-access-7nmph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:10:52 crc kubenswrapper[4706]: I1206 14:10:52.518385 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nmph\" (UniqueName: \"kubernetes.io/projected/9160e5ea-636c-4924-8ba1-1dfd13dfa0cc-kube-api-access-7nmph\") on node \"crc\" DevicePath \"\"" Dec 06 14:10:52 crc kubenswrapper[4706]: I1206 14:10:52.973950 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-6xsvs" event={"ID":"352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9","Type":"ContainerStarted","Data":"5f39d0295437686b34ff846dc9258109e8568381c6a48e7cb5a202f60d83f2c6"} Dec 06 14:10:52 crc kubenswrapper[4706]: I1206 14:10:52.974358 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-6xsvs" event={"ID":"352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9","Type":"ContainerStarted","Data":"ba7e718b08d89cc9658005cba4c6f43ef4e18764de6a25825d7eb3c55f5dc0d9"} Dec 06 14:10:52 crc kubenswrapper[4706]: I1206 14:10:52.976517 4706 generic.go:334] "Generic (PLEG): container finished" podID="9160e5ea-636c-4924-8ba1-1dfd13dfa0cc" containerID="7f1b924815e7d644d23cbb802884fe1a940f8ecc413e11e0ee7a3f6ddf434af7" exitCode=0 Dec 06 14:10:52 crc kubenswrapper[4706]: I1206 14:10:52.976586 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-m4592" event={"ID":"9160e5ea-636c-4924-8ba1-1dfd13dfa0cc","Type":"ContainerDied","Data":"7f1b924815e7d644d23cbb802884fe1a940f8ecc413e11e0ee7a3f6ddf434af7"} Dec 06 14:10:52 crc kubenswrapper[4706]: I1206 14:10:52.976627 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-m4592" event={"ID":"9160e5ea-636c-4924-8ba1-1dfd13dfa0cc","Type":"ContainerDied","Data":"054abdabb0c0f696fc5c971b9b1acf846fafaefedaf776834798bb36fa31f47d"} Dec 06 14:10:52 crc kubenswrapper[4706]: I1206 14:10:52.976619 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-m4592" Dec 06 14:10:52 crc kubenswrapper[4706]: I1206 14:10:52.976698 4706 scope.go:117] "RemoveContainer" containerID="7f1b924815e7d644d23cbb802884fe1a940f8ecc413e11e0ee7a3f6ddf434af7" Dec 06 14:10:52 crc kubenswrapper[4706]: I1206 14:10:52.997416 4706 scope.go:117] "RemoveContainer" containerID="7f1b924815e7d644d23cbb802884fe1a940f8ecc413e11e0ee7a3f6ddf434af7" Dec 06 14:10:52 crc kubenswrapper[4706]: E1206 14:10:52.998278 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f1b924815e7d644d23cbb802884fe1a940f8ecc413e11e0ee7a3f6ddf434af7\": container with ID starting with 7f1b924815e7d644d23cbb802884fe1a940f8ecc413e11e0ee7a3f6ddf434af7 not found: ID does not exist" containerID="7f1b924815e7d644d23cbb802884fe1a940f8ecc413e11e0ee7a3f6ddf434af7" Dec 06 14:10:52 crc kubenswrapper[4706]: I1206 14:10:52.998335 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f1b924815e7d644d23cbb802884fe1a940f8ecc413e11e0ee7a3f6ddf434af7"} err="failed to get container status \"7f1b924815e7d644d23cbb802884fe1a940f8ecc413e11e0ee7a3f6ddf434af7\": rpc error: code = NotFound desc = could not find container \"7f1b924815e7d644d23cbb802884fe1a940f8ecc413e11e0ee7a3f6ddf434af7\": container with ID starting with 7f1b924815e7d644d23cbb802884fe1a940f8ecc413e11e0ee7a3f6ddf434af7 not found: ID does not exist" Dec 06 14:10:53 crc kubenswrapper[4706]: I1206 14:10:53.006661 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-6xsvs" podStartSLOduration=1.584434141 podStartE2EDuration="2.006650267s" podCreationTimestamp="2025-12-06 14:10:51 +0000 UTC" firstStartedPulling="2025-12-06 14:10:52.246420926 +0000 UTC m=+895.082213783" lastFinishedPulling="2025-12-06 14:10:52.668637052 +0000 UTC m=+895.504429909" observedRunningTime="2025-12-06 14:10:52.990999592 +0000 UTC m=+895.826792449" watchObservedRunningTime="2025-12-06 14:10:53.006650267 +0000 UTC m=+895.842443124" Dec 06 14:10:53 crc kubenswrapper[4706]: I1206 14:10:53.011353 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-m4592"] Dec 06 14:10:53 crc kubenswrapper[4706]: I1206 14:10:53.015077 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-m4592"] Dec 06 14:10:53 crc kubenswrapper[4706]: I1206 14:10:53.525075 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9160e5ea-636c-4924-8ba1-1dfd13dfa0cc" path="/var/lib/kubelet/pods/9160e5ea-636c-4924-8ba1-1dfd13dfa0cc/volumes" Dec 06 14:10:56 crc kubenswrapper[4706]: I1206 14:10:56.038577 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/memcached-0" Dec 06 14:11:01 crc kubenswrapper[4706]: I1206 14:11:01.790553 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-6xsvs" Dec 06 14:11:01 crc kubenswrapper[4706]: I1206 14:11:01.791022 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-6xsvs" Dec 06 14:11:01 crc kubenswrapper[4706]: I1206 14:11:01.828018 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-6xsvs" Dec 06 14:11:02 crc kubenswrapper[4706]: I1206 14:11:02.094520 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-6xsvs" Dec 06 14:11:05 crc kubenswrapper[4706]: I1206 14:11:05.065566 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:11:05 crc kubenswrapper[4706]: I1206 14:11:05.184648 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:11:07 crc kubenswrapper[4706]: I1206 14:11:07.101785 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:11:07 crc kubenswrapper[4706]: I1206 14:11:07.181225 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:11:09 crc kubenswrapper[4706]: I1206 14:11:09.358891 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx"] Dec 06 14:11:09 crc kubenswrapper[4706]: E1206 14:11:09.359557 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9160e5ea-636c-4924-8ba1-1dfd13dfa0cc" containerName="registry-server" Dec 06 14:11:09 crc kubenswrapper[4706]: I1206 14:11:09.359594 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="9160e5ea-636c-4924-8ba1-1dfd13dfa0cc" containerName="registry-server" Dec 06 14:11:09 crc kubenswrapper[4706]: I1206 14:11:09.359777 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="9160e5ea-636c-4924-8ba1-1dfd13dfa0cc" containerName="registry-server" Dec 06 14:11:09 crc kubenswrapper[4706]: I1206 14:11:09.361289 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" Dec 06 14:11:09 crc kubenswrapper[4706]: I1206 14:11:09.363199 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-fc6v2" Dec 06 14:11:09 crc kubenswrapper[4706]: I1206 14:11:09.371299 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx"] Dec 06 14:11:09 crc kubenswrapper[4706]: I1206 14:11:09.449697 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxh8l\" (UniqueName: \"kubernetes.io/projected/380fba54-3c03-4e8f-b702-6a2ad96c2744-kube-api-access-bxh8l\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx\" (UID: \"380fba54-3c03-4e8f-b702-6a2ad96c2744\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" Dec 06 14:11:09 crc kubenswrapper[4706]: I1206 14:11:09.449771 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/380fba54-3c03-4e8f-b702-6a2ad96c2744-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx\" (UID: \"380fba54-3c03-4e8f-b702-6a2ad96c2744\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" Dec 06 14:11:09 crc kubenswrapper[4706]: I1206 14:11:09.449944 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/380fba54-3c03-4e8f-b702-6a2ad96c2744-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx\" (UID: \"380fba54-3c03-4e8f-b702-6a2ad96c2744\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" Dec 06 14:11:09 crc kubenswrapper[4706]: I1206 14:11:09.551143 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxh8l\" (UniqueName: \"kubernetes.io/projected/380fba54-3c03-4e8f-b702-6a2ad96c2744-kube-api-access-bxh8l\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx\" (UID: \"380fba54-3c03-4e8f-b702-6a2ad96c2744\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" Dec 06 14:11:09 crc kubenswrapper[4706]: I1206 14:11:09.551219 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/380fba54-3c03-4e8f-b702-6a2ad96c2744-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx\" (UID: \"380fba54-3c03-4e8f-b702-6a2ad96c2744\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" Dec 06 14:11:09 crc kubenswrapper[4706]: I1206 14:11:09.551280 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/380fba54-3c03-4e8f-b702-6a2ad96c2744-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx\" (UID: \"380fba54-3c03-4e8f-b702-6a2ad96c2744\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" Dec 06 14:11:09 crc kubenswrapper[4706]: I1206 14:11:09.551835 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/380fba54-3c03-4e8f-b702-6a2ad96c2744-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx\" (UID: \"380fba54-3c03-4e8f-b702-6a2ad96c2744\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" Dec 06 14:11:09 crc kubenswrapper[4706]: I1206 14:11:09.551938 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/380fba54-3c03-4e8f-b702-6a2ad96c2744-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx\" (UID: \"380fba54-3c03-4e8f-b702-6a2ad96c2744\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" Dec 06 14:11:09 crc kubenswrapper[4706]: I1206 14:11:09.576152 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxh8l\" (UniqueName: \"kubernetes.io/projected/380fba54-3c03-4e8f-b702-6a2ad96c2744-kube-api-access-bxh8l\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx\" (UID: \"380fba54-3c03-4e8f-b702-6a2ad96c2744\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" Dec 06 14:11:09 crc kubenswrapper[4706]: I1206 14:11:09.697775 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" Dec 06 14:11:09 crc kubenswrapper[4706]: I1206 14:11:09.973304 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx"] Dec 06 14:11:10 crc kubenswrapper[4706]: I1206 14:11:10.128523 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" event={"ID":"380fba54-3c03-4e8f-b702-6a2ad96c2744","Type":"ContainerStarted","Data":"80164252d1e74dde40b1bb3dcdace71bdeb67e33fcdf568fe0a4dc32e6631c46"} Dec 06 14:11:11 crc kubenswrapper[4706]: I1206 14:11:11.139842 4706 generic.go:334] "Generic (PLEG): container finished" podID="380fba54-3c03-4e8f-b702-6a2ad96c2744" containerID="e76355becb4e03558d1bb2f1a40134e7331e1c0f23b58bd50a00061f90f0f64f" exitCode=0 Dec 06 14:11:11 crc kubenswrapper[4706]: I1206 14:11:11.139942 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" event={"ID":"380fba54-3c03-4e8f-b702-6a2ad96c2744","Type":"ContainerDied","Data":"e76355becb4e03558d1bb2f1a40134e7331e1c0f23b58bd50a00061f90f0f64f"} Dec 06 14:11:12 crc kubenswrapper[4706]: I1206 14:11:12.152740 4706 generic.go:334] "Generic (PLEG): container finished" podID="380fba54-3c03-4e8f-b702-6a2ad96c2744" containerID="74029b0dde685b98feafe6433cf565edfcc2b8cc0fb755106524a513baa7d500" exitCode=0 Dec 06 14:11:12 crc kubenswrapper[4706]: I1206 14:11:12.152811 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" event={"ID":"380fba54-3c03-4e8f-b702-6a2ad96c2744","Type":"ContainerDied","Data":"74029b0dde685b98feafe6433cf565edfcc2b8cc0fb755106524a513baa7d500"} Dec 06 14:11:13 crc kubenswrapper[4706]: I1206 14:11:13.165098 4706 generic.go:334] "Generic (PLEG): container finished" podID="380fba54-3c03-4e8f-b702-6a2ad96c2744" containerID="fba5757c3407cbacde9f45f2a39ffdd2c94fb9a6da8a4a223483c07c14dbd10e" exitCode=0 Dec 06 14:11:13 crc kubenswrapper[4706]: I1206 14:11:13.165159 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" event={"ID":"380fba54-3c03-4e8f-b702-6a2ad96c2744","Type":"ContainerDied","Data":"fba5757c3407cbacde9f45f2a39ffdd2c94fb9a6da8a4a223483c07c14dbd10e"} Dec 06 14:11:14 crc kubenswrapper[4706]: I1206 14:11:14.563395 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" Dec 06 14:11:14 crc kubenswrapper[4706]: I1206 14:11:14.635908 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/380fba54-3c03-4e8f-b702-6a2ad96c2744-bundle\") pod \"380fba54-3c03-4e8f-b702-6a2ad96c2744\" (UID: \"380fba54-3c03-4e8f-b702-6a2ad96c2744\") " Dec 06 14:11:14 crc kubenswrapper[4706]: I1206 14:11:14.636002 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxh8l\" (UniqueName: \"kubernetes.io/projected/380fba54-3c03-4e8f-b702-6a2ad96c2744-kube-api-access-bxh8l\") pod \"380fba54-3c03-4e8f-b702-6a2ad96c2744\" (UID: \"380fba54-3c03-4e8f-b702-6a2ad96c2744\") " Dec 06 14:11:14 crc kubenswrapper[4706]: I1206 14:11:14.636076 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/380fba54-3c03-4e8f-b702-6a2ad96c2744-util\") pod \"380fba54-3c03-4e8f-b702-6a2ad96c2744\" (UID: \"380fba54-3c03-4e8f-b702-6a2ad96c2744\") " Dec 06 14:11:14 crc kubenswrapper[4706]: I1206 14:11:14.637155 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/380fba54-3c03-4e8f-b702-6a2ad96c2744-bundle" (OuterVolumeSpecName: "bundle") pod "380fba54-3c03-4e8f-b702-6a2ad96c2744" (UID: "380fba54-3c03-4e8f-b702-6a2ad96c2744"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:11:14 crc kubenswrapper[4706]: I1206 14:11:14.649238 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/380fba54-3c03-4e8f-b702-6a2ad96c2744-kube-api-access-bxh8l" (OuterVolumeSpecName: "kube-api-access-bxh8l") pod "380fba54-3c03-4e8f-b702-6a2ad96c2744" (UID: "380fba54-3c03-4e8f-b702-6a2ad96c2744"). InnerVolumeSpecName "kube-api-access-bxh8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:11:14 crc kubenswrapper[4706]: I1206 14:11:14.656631 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/380fba54-3c03-4e8f-b702-6a2ad96c2744-util" (OuterVolumeSpecName: "util") pod "380fba54-3c03-4e8f-b702-6a2ad96c2744" (UID: "380fba54-3c03-4e8f-b702-6a2ad96c2744"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:11:14 crc kubenswrapper[4706]: I1206 14:11:14.737642 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxh8l\" (UniqueName: \"kubernetes.io/projected/380fba54-3c03-4e8f-b702-6a2ad96c2744-kube-api-access-bxh8l\") on node \"crc\" DevicePath \"\"" Dec 06 14:11:14 crc kubenswrapper[4706]: I1206 14:11:14.737692 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/380fba54-3c03-4e8f-b702-6a2ad96c2744-util\") on node \"crc\" DevicePath \"\"" Dec 06 14:11:14 crc kubenswrapper[4706]: I1206 14:11:14.737710 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/380fba54-3c03-4e8f-b702-6a2ad96c2744-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 14:11:15 crc kubenswrapper[4706]: I1206 14:11:15.185802 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" event={"ID":"380fba54-3c03-4e8f-b702-6a2ad96c2744","Type":"ContainerDied","Data":"80164252d1e74dde40b1bb3dcdace71bdeb67e33fcdf568fe0a4dc32e6631c46"} Dec 06 14:11:15 crc kubenswrapper[4706]: I1206 14:11:15.185850 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80164252d1e74dde40b1bb3dcdace71bdeb67e33fcdf568fe0a4dc32e6631c46" Dec 06 14:11:15 crc kubenswrapper[4706]: I1206 14:11:15.185923 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx" Dec 06 14:11:22 crc kubenswrapper[4706]: I1206 14:11:22.942815 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp"] Dec 06 14:11:22 crc kubenswrapper[4706]: E1206 14:11:22.943724 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="380fba54-3c03-4e8f-b702-6a2ad96c2744" containerName="pull" Dec 06 14:11:22 crc kubenswrapper[4706]: I1206 14:11:22.943745 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="380fba54-3c03-4e8f-b702-6a2ad96c2744" containerName="pull" Dec 06 14:11:22 crc kubenswrapper[4706]: E1206 14:11:22.943768 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="380fba54-3c03-4e8f-b702-6a2ad96c2744" containerName="extract" Dec 06 14:11:22 crc kubenswrapper[4706]: I1206 14:11:22.943781 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="380fba54-3c03-4e8f-b702-6a2ad96c2744" containerName="extract" Dec 06 14:11:22 crc kubenswrapper[4706]: E1206 14:11:22.943822 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="380fba54-3c03-4e8f-b702-6a2ad96c2744" containerName="util" Dec 06 14:11:22 crc kubenswrapper[4706]: I1206 14:11:22.943837 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="380fba54-3c03-4e8f-b702-6a2ad96c2744" containerName="util" Dec 06 14:11:22 crc kubenswrapper[4706]: I1206 14:11:22.944084 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="380fba54-3c03-4e8f-b702-6a2ad96c2744" containerName="extract" Dec 06 14:11:22 crc kubenswrapper[4706]: I1206 14:11:22.944731 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp" Dec 06 14:11:22 crc kubenswrapper[4706]: I1206 14:11:22.957268 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-6q47s" Dec 06 14:11:22 crc kubenswrapper[4706]: I1206 14:11:22.981464 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp"] Dec 06 14:11:23 crc kubenswrapper[4706]: I1206 14:11:23.057852 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k62b8\" (UniqueName: \"kubernetes.io/projected/18cc7358-0421-415c-a37b-6dba9a764d93-kube-api-access-k62b8\") pod \"rabbitmq-cluster-operator-779fc9694b-pqwvp\" (UID: \"18cc7358-0421-415c-a37b-6dba9a764d93\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp" Dec 06 14:11:23 crc kubenswrapper[4706]: I1206 14:11:23.159525 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k62b8\" (UniqueName: \"kubernetes.io/projected/18cc7358-0421-415c-a37b-6dba9a764d93-kube-api-access-k62b8\") pod \"rabbitmq-cluster-operator-779fc9694b-pqwvp\" (UID: \"18cc7358-0421-415c-a37b-6dba9a764d93\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp" Dec 06 14:11:23 crc kubenswrapper[4706]: I1206 14:11:23.186614 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k62b8\" (UniqueName: \"kubernetes.io/projected/18cc7358-0421-415c-a37b-6dba9a764d93-kube-api-access-k62b8\") pod \"rabbitmq-cluster-operator-779fc9694b-pqwvp\" (UID: \"18cc7358-0421-415c-a37b-6dba9a764d93\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp" Dec 06 14:11:23 crc kubenswrapper[4706]: I1206 14:11:23.279133 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp" Dec 06 14:11:23 crc kubenswrapper[4706]: I1206 14:11:23.532787 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp"] Dec 06 14:11:24 crc kubenswrapper[4706]: I1206 14:11:24.256279 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp" event={"ID":"18cc7358-0421-415c-a37b-6dba9a764d93","Type":"ContainerStarted","Data":"e9e2f744285d2bf82711a97c6c946ac701f978c36ffd90210a9d751dbffd9d38"} Dec 06 14:11:27 crc kubenswrapper[4706]: I1206 14:11:27.281281 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp" event={"ID":"18cc7358-0421-415c-a37b-6dba9a764d93","Type":"ContainerStarted","Data":"1779e8484d779e98a1cc93c6c5788c8cb6a4a5645fb3ac2d27b558da926da210"} Dec 06 14:11:27 crc kubenswrapper[4706]: I1206 14:11:27.304630 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp" podStartSLOduration=1.8039531260000001 podStartE2EDuration="5.304604609s" podCreationTimestamp="2025-12-06 14:11:22 +0000 UTC" firstStartedPulling="2025-12-06 14:11:23.53763568 +0000 UTC m=+926.373428537" lastFinishedPulling="2025-12-06 14:11:27.038287163 +0000 UTC m=+929.874080020" observedRunningTime="2025-12-06 14:11:27.298739793 +0000 UTC m=+930.134532710" watchObservedRunningTime="2025-12-06 14:11:27.304604609 +0000 UTC m=+930.140397466" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.214132 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.215811 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.219097 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"rabbitmq-erlang-cookie" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.219684 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"rabbitmq-plugins-conf" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.220445 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"rabbitmq-server-dockercfg-7pp8x" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.220800 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"rabbitmq-default-user" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.222415 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"rabbitmq-server-conf" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.241911 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.312814 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1a988e02-edca-43b9-b5e5-92410a885971\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a988e02-edca-43b9-b5e5-92410a885971\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.312970 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85dmj\" (UniqueName: \"kubernetes.io/projected/23f8c907-7a56-4da3-aefb-929a43497eb2-kube-api-access-85dmj\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.313070 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.313152 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/23f8c907-7a56-4da3-aefb-929a43497eb2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.313236 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.313303 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/23f8c907-7a56-4da3-aefb-929a43497eb2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.313391 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/23f8c907-7a56-4da3-aefb-929a43497eb2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.313542 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.415439 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.415560 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/23f8c907-7a56-4da3-aefb-929a43497eb2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.415618 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/23f8c907-7a56-4da3-aefb-929a43497eb2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.415689 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.415754 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1a988e02-edca-43b9-b5e5-92410a885971\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a988e02-edca-43b9-b5e5-92410a885971\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.415864 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85dmj\" (UniqueName: \"kubernetes.io/projected/23f8c907-7a56-4da3-aefb-929a43497eb2-kube-api-access-85dmj\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.415954 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.416000 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/23f8c907-7a56-4da3-aefb-929a43497eb2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.417188 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.417536 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.418216 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/23f8c907-7a56-4da3-aefb-929a43497eb2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.422742 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/23f8c907-7a56-4da3-aefb-929a43497eb2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.424423 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/23f8c907-7a56-4da3-aefb-929a43497eb2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.424902 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.424956 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1a988e02-edca-43b9-b5e5-92410a885971\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a988e02-edca-43b9-b5e5-92410a885971\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/84997b4c73a1ad2b8d9a9b8eb84f47618d4409ddf872c9fbc4de6d71235c5253/globalmount\"" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.439730 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.441438 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85dmj\" (UniqueName: \"kubernetes.io/projected/23f8c907-7a56-4da3-aefb-929a43497eb2-kube-api-access-85dmj\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.451047 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1a988e02-edca-43b9-b5e5-92410a885971\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a988e02-edca-43b9-b5e5-92410a885971\") pod \"rabbitmq-server-0\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.550422 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:11:33 crc kubenswrapper[4706]: I1206 14:11:33.820820 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 06 14:11:34 crc kubenswrapper[4706]: I1206 14:11:34.334527 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"23f8c907-7a56-4da3-aefb-929a43497eb2","Type":"ContainerStarted","Data":"958295388ed3710103707281add52f4b201f078a5e63d4377f3512e64b08714f"} Dec 06 14:11:34 crc kubenswrapper[4706]: I1206 14:11:34.872215 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-nhn8p"] Dec 06 14:11:34 crc kubenswrapper[4706]: I1206 14:11:34.873510 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-nhn8p" Dec 06 14:11:34 crc kubenswrapper[4706]: I1206 14:11:34.876564 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-m6nct" Dec 06 14:11:34 crc kubenswrapper[4706]: I1206 14:11:34.890876 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-nhn8p"] Dec 06 14:11:34 crc kubenswrapper[4706]: I1206 14:11:34.940377 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8dgb\" (UniqueName: \"kubernetes.io/projected/d5149dd3-dea6-4c73-b75e-5f41f3c57420-kube-api-access-v8dgb\") pod \"keystone-operator-index-nhn8p\" (UID: \"d5149dd3-dea6-4c73-b75e-5f41f3c57420\") " pod="openstack-operators/keystone-operator-index-nhn8p" Dec 06 14:11:35 crc kubenswrapper[4706]: I1206 14:11:35.046420 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8dgb\" (UniqueName: \"kubernetes.io/projected/d5149dd3-dea6-4c73-b75e-5f41f3c57420-kube-api-access-v8dgb\") pod \"keystone-operator-index-nhn8p\" (UID: \"d5149dd3-dea6-4c73-b75e-5f41f3c57420\") " pod="openstack-operators/keystone-operator-index-nhn8p" Dec 06 14:11:35 crc kubenswrapper[4706]: I1206 14:11:35.077586 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8dgb\" (UniqueName: \"kubernetes.io/projected/d5149dd3-dea6-4c73-b75e-5f41f3c57420-kube-api-access-v8dgb\") pod \"keystone-operator-index-nhn8p\" (UID: \"d5149dd3-dea6-4c73-b75e-5f41f3c57420\") " pod="openstack-operators/keystone-operator-index-nhn8p" Dec 06 14:11:35 crc kubenswrapper[4706]: I1206 14:11:35.214215 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-nhn8p" Dec 06 14:11:35 crc kubenswrapper[4706]: I1206 14:11:35.408079 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-nhn8p"] Dec 06 14:11:35 crc kubenswrapper[4706]: W1206 14:11:35.417574 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5149dd3_dea6_4c73_b75e_5f41f3c57420.slice/crio-7e531f65465d6e22c07e2e6693b3d39c85e653b4cfbd65ac8af6e8f7d326740b WatchSource:0}: Error finding container 7e531f65465d6e22c07e2e6693b3d39c85e653b4cfbd65ac8af6e8f7d326740b: Status 404 returned error can't find the container with id 7e531f65465d6e22c07e2e6693b3d39c85e653b4cfbd65ac8af6e8f7d326740b Dec 06 14:11:36 crc kubenswrapper[4706]: I1206 14:11:36.360777 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-nhn8p" event={"ID":"d5149dd3-dea6-4c73-b75e-5f41f3c57420","Type":"ContainerStarted","Data":"7e531f65465d6e22c07e2e6693b3d39c85e653b4cfbd65ac8af6e8f7d326740b"} Dec 06 14:11:39 crc kubenswrapper[4706]: I1206 14:11:39.253011 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-nhn8p"] Dec 06 14:11:39 crc kubenswrapper[4706]: I1206 14:11:39.863423 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-z4lsq"] Dec 06 14:11:39 crc kubenswrapper[4706]: I1206 14:11:39.864252 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-z4lsq" Dec 06 14:11:39 crc kubenswrapper[4706]: I1206 14:11:39.885056 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-z4lsq"] Dec 06 14:11:39 crc kubenswrapper[4706]: I1206 14:11:39.926960 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8rw5\" (UniqueName: \"kubernetes.io/projected/7490dd58-b444-46c5-ad33-b953af4e6617-kube-api-access-p8rw5\") pod \"keystone-operator-index-z4lsq\" (UID: \"7490dd58-b444-46c5-ad33-b953af4e6617\") " pod="openstack-operators/keystone-operator-index-z4lsq" Dec 06 14:11:40 crc kubenswrapper[4706]: I1206 14:11:40.028503 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8rw5\" (UniqueName: \"kubernetes.io/projected/7490dd58-b444-46c5-ad33-b953af4e6617-kube-api-access-p8rw5\") pod \"keystone-operator-index-z4lsq\" (UID: \"7490dd58-b444-46c5-ad33-b953af4e6617\") " pod="openstack-operators/keystone-operator-index-z4lsq" Dec 06 14:11:40 crc kubenswrapper[4706]: I1206 14:11:40.050694 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8rw5\" (UniqueName: \"kubernetes.io/projected/7490dd58-b444-46c5-ad33-b953af4e6617-kube-api-access-p8rw5\") pod \"keystone-operator-index-z4lsq\" (UID: \"7490dd58-b444-46c5-ad33-b953af4e6617\") " pod="openstack-operators/keystone-operator-index-z4lsq" Dec 06 14:11:40 crc kubenswrapper[4706]: I1206 14:11:40.221864 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-z4lsq" Dec 06 14:11:42 crc kubenswrapper[4706]: I1206 14:11:42.879233 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-z4lsq"] Dec 06 14:11:42 crc kubenswrapper[4706]: W1206 14:11:42.883271 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7490dd58_b444_46c5_ad33_b953af4e6617.slice/crio-c40218a5fa31108baaef985e772ec444035a71eef14ef91a45e74bceca360f1f WatchSource:0}: Error finding container c40218a5fa31108baaef985e772ec444035a71eef14ef91a45e74bceca360f1f: Status 404 returned error can't find the container with id c40218a5fa31108baaef985e772ec444035a71eef14ef91a45e74bceca360f1f Dec 06 14:11:43 crc kubenswrapper[4706]: I1206 14:11:43.418185 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-z4lsq" event={"ID":"7490dd58-b444-46c5-ad33-b953af4e6617","Type":"ContainerStarted","Data":"c40218a5fa31108baaef985e772ec444035a71eef14ef91a45e74bceca360f1f"} Dec 06 14:11:43 crc kubenswrapper[4706]: I1206 14:11:43.420747 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-nhn8p" event={"ID":"d5149dd3-dea6-4c73-b75e-5f41f3c57420","Type":"ContainerStarted","Data":"c21a35b289f0511efefdc65963592b0fdb07e347fb5eddb52be553960a02fe10"} Dec 06 14:11:43 crc kubenswrapper[4706]: I1206 14:11:43.420903 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-nhn8p" podUID="d5149dd3-dea6-4c73-b75e-5f41f3c57420" containerName="registry-server" containerID="cri-o://c21a35b289f0511efefdc65963592b0fdb07e347fb5eddb52be553960a02fe10" gracePeriod=2 Dec 06 14:11:43 crc kubenswrapper[4706]: I1206 14:11:43.446256 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-nhn8p" podStartSLOduration=2.240374325 podStartE2EDuration="9.446230512s" podCreationTimestamp="2025-12-06 14:11:34 +0000 UTC" firstStartedPulling="2025-12-06 14:11:35.420565646 +0000 UTC m=+938.256358503" lastFinishedPulling="2025-12-06 14:11:42.626421823 +0000 UTC m=+945.462214690" observedRunningTime="2025-12-06 14:11:43.442144774 +0000 UTC m=+946.277937661" watchObservedRunningTime="2025-12-06 14:11:43.446230512 +0000 UTC m=+946.282023409" Dec 06 14:11:43 crc kubenswrapper[4706]: I1206 14:11:43.848486 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-nhn8p" Dec 06 14:11:44 crc kubenswrapper[4706]: I1206 14:11:44.036938 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8dgb\" (UniqueName: \"kubernetes.io/projected/d5149dd3-dea6-4c73-b75e-5f41f3c57420-kube-api-access-v8dgb\") pod \"d5149dd3-dea6-4c73-b75e-5f41f3c57420\" (UID: \"d5149dd3-dea6-4c73-b75e-5f41f3c57420\") " Dec 06 14:11:44 crc kubenswrapper[4706]: I1206 14:11:44.052754 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5149dd3-dea6-4c73-b75e-5f41f3c57420-kube-api-access-v8dgb" (OuterVolumeSpecName: "kube-api-access-v8dgb") pod "d5149dd3-dea6-4c73-b75e-5f41f3c57420" (UID: "d5149dd3-dea6-4c73-b75e-5f41f3c57420"). InnerVolumeSpecName "kube-api-access-v8dgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:11:44 crc kubenswrapper[4706]: I1206 14:11:44.138793 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8dgb\" (UniqueName: \"kubernetes.io/projected/d5149dd3-dea6-4c73-b75e-5f41f3c57420-kube-api-access-v8dgb\") on node \"crc\" DevicePath \"\"" Dec 06 14:11:44 crc kubenswrapper[4706]: I1206 14:11:44.431692 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-z4lsq" event={"ID":"7490dd58-b444-46c5-ad33-b953af4e6617","Type":"ContainerStarted","Data":"e6b2578eedd50f1964b40e7247cfb412e66462b9860027502721d0a39f08c826"} Dec 06 14:11:44 crc kubenswrapper[4706]: I1206 14:11:44.434459 4706 generic.go:334] "Generic (PLEG): container finished" podID="d5149dd3-dea6-4c73-b75e-5f41f3c57420" containerID="c21a35b289f0511efefdc65963592b0fdb07e347fb5eddb52be553960a02fe10" exitCode=0 Dec 06 14:11:44 crc kubenswrapper[4706]: I1206 14:11:44.434556 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-nhn8p" event={"ID":"d5149dd3-dea6-4c73-b75e-5f41f3c57420","Type":"ContainerDied","Data":"c21a35b289f0511efefdc65963592b0fdb07e347fb5eddb52be553960a02fe10"} Dec 06 14:11:44 crc kubenswrapper[4706]: I1206 14:11:44.434599 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-nhn8p" event={"ID":"d5149dd3-dea6-4c73-b75e-5f41f3c57420","Type":"ContainerDied","Data":"7e531f65465d6e22c07e2e6693b3d39c85e653b4cfbd65ac8af6e8f7d326740b"} Dec 06 14:11:44 crc kubenswrapper[4706]: I1206 14:11:44.434634 4706 scope.go:117] "RemoveContainer" containerID="c21a35b289f0511efefdc65963592b0fdb07e347fb5eddb52be553960a02fe10" Dec 06 14:11:44 crc kubenswrapper[4706]: I1206 14:11:44.434820 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-nhn8p" Dec 06 14:11:44 crc kubenswrapper[4706]: I1206 14:11:44.436928 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"23f8c907-7a56-4da3-aefb-929a43497eb2","Type":"ContainerStarted","Data":"6a2c055976a1f626f8c484f7e2eab83c12e64177e0a91754b946043503eb7531"} Dec 06 14:11:44 crc kubenswrapper[4706]: I1206 14:11:44.463799 4706 scope.go:117] "RemoveContainer" containerID="c21a35b289f0511efefdc65963592b0fdb07e347fb5eddb52be553960a02fe10" Dec 06 14:11:44 crc kubenswrapper[4706]: E1206 14:11:44.464158 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c21a35b289f0511efefdc65963592b0fdb07e347fb5eddb52be553960a02fe10\": container with ID starting with c21a35b289f0511efefdc65963592b0fdb07e347fb5eddb52be553960a02fe10 not found: ID does not exist" containerID="c21a35b289f0511efefdc65963592b0fdb07e347fb5eddb52be553960a02fe10" Dec 06 14:11:44 crc kubenswrapper[4706]: I1206 14:11:44.464196 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c21a35b289f0511efefdc65963592b0fdb07e347fb5eddb52be553960a02fe10"} err="failed to get container status \"c21a35b289f0511efefdc65963592b0fdb07e347fb5eddb52be553960a02fe10\": rpc error: code = NotFound desc = could not find container \"c21a35b289f0511efefdc65963592b0fdb07e347fb5eddb52be553960a02fe10\": container with ID starting with c21a35b289f0511efefdc65963592b0fdb07e347fb5eddb52be553960a02fe10 not found: ID does not exist" Dec 06 14:11:44 crc kubenswrapper[4706]: I1206 14:11:44.466618 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-z4lsq" podStartSLOduration=5.043810903 podStartE2EDuration="5.466591064s" podCreationTimestamp="2025-12-06 14:11:39 +0000 UTC" firstStartedPulling="2025-12-06 14:11:42.88751625 +0000 UTC m=+945.723309127" lastFinishedPulling="2025-12-06 14:11:43.310296391 +0000 UTC m=+946.146089288" observedRunningTime="2025-12-06 14:11:44.457022111 +0000 UTC m=+947.292814978" watchObservedRunningTime="2025-12-06 14:11:44.466591064 +0000 UTC m=+947.302383961" Dec 06 14:11:44 crc kubenswrapper[4706]: I1206 14:11:44.507374 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-nhn8p"] Dec 06 14:11:44 crc kubenswrapper[4706]: I1206 14:11:44.511614 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-nhn8p"] Dec 06 14:11:44 crc kubenswrapper[4706]: I1206 14:11:44.580273 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:11:44 crc kubenswrapper[4706]: I1206 14:11:44.580371 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:11:45 crc kubenswrapper[4706]: I1206 14:11:45.524868 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5149dd3-dea6-4c73-b75e-5f41f3c57420" path="/var/lib/kubelet/pods/d5149dd3-dea6-4c73-b75e-5f41f3c57420/volumes" Dec 06 14:11:50 crc kubenswrapper[4706]: I1206 14:11:50.222858 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-z4lsq" Dec 06 14:11:50 crc kubenswrapper[4706]: I1206 14:11:50.223385 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-z4lsq" Dec 06 14:11:50 crc kubenswrapper[4706]: I1206 14:11:50.265952 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-z4lsq" Dec 06 14:11:50 crc kubenswrapper[4706]: I1206 14:11:50.522541 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-z4lsq" Dec 06 14:11:52 crc kubenswrapper[4706]: I1206 14:11:52.675444 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qtn5m"] Dec 06 14:11:52 crc kubenswrapper[4706]: E1206 14:11:52.676174 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5149dd3-dea6-4c73-b75e-5f41f3c57420" containerName="registry-server" Dec 06 14:11:52 crc kubenswrapper[4706]: I1206 14:11:52.676196 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5149dd3-dea6-4c73-b75e-5f41f3c57420" containerName="registry-server" Dec 06 14:11:52 crc kubenswrapper[4706]: I1206 14:11:52.676421 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5149dd3-dea6-4c73-b75e-5f41f3c57420" containerName="registry-server" Dec 06 14:11:52 crc kubenswrapper[4706]: I1206 14:11:52.677791 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qtn5m" Dec 06 14:11:52 crc kubenswrapper[4706]: I1206 14:11:52.687498 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtn5m"] Dec 06 14:11:52 crc kubenswrapper[4706]: I1206 14:11:52.773845 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp484\" (UniqueName: \"kubernetes.io/projected/a055d801-0918-4b3b-91b9-a311d58f9d51-kube-api-access-dp484\") pod \"redhat-marketplace-qtn5m\" (UID: \"a055d801-0918-4b3b-91b9-a311d58f9d51\") " pod="openshift-marketplace/redhat-marketplace-qtn5m" Dec 06 14:11:52 crc kubenswrapper[4706]: I1206 14:11:52.773950 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a055d801-0918-4b3b-91b9-a311d58f9d51-utilities\") pod \"redhat-marketplace-qtn5m\" (UID: \"a055d801-0918-4b3b-91b9-a311d58f9d51\") " pod="openshift-marketplace/redhat-marketplace-qtn5m" Dec 06 14:11:52 crc kubenswrapper[4706]: I1206 14:11:52.774030 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a055d801-0918-4b3b-91b9-a311d58f9d51-catalog-content\") pod \"redhat-marketplace-qtn5m\" (UID: \"a055d801-0918-4b3b-91b9-a311d58f9d51\") " pod="openshift-marketplace/redhat-marketplace-qtn5m" Dec 06 14:11:52 crc kubenswrapper[4706]: I1206 14:11:52.876220 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a055d801-0918-4b3b-91b9-a311d58f9d51-utilities\") pod \"redhat-marketplace-qtn5m\" (UID: \"a055d801-0918-4b3b-91b9-a311d58f9d51\") " pod="openshift-marketplace/redhat-marketplace-qtn5m" Dec 06 14:11:52 crc kubenswrapper[4706]: I1206 14:11:52.876442 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a055d801-0918-4b3b-91b9-a311d58f9d51-catalog-content\") pod \"redhat-marketplace-qtn5m\" (UID: \"a055d801-0918-4b3b-91b9-a311d58f9d51\") " pod="openshift-marketplace/redhat-marketplace-qtn5m" Dec 06 14:11:52 crc kubenswrapper[4706]: I1206 14:11:52.876628 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp484\" (UniqueName: \"kubernetes.io/projected/a055d801-0918-4b3b-91b9-a311d58f9d51-kube-api-access-dp484\") pod \"redhat-marketplace-qtn5m\" (UID: \"a055d801-0918-4b3b-91b9-a311d58f9d51\") " pod="openshift-marketplace/redhat-marketplace-qtn5m" Dec 06 14:11:52 crc kubenswrapper[4706]: I1206 14:11:52.876971 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a055d801-0918-4b3b-91b9-a311d58f9d51-utilities\") pod \"redhat-marketplace-qtn5m\" (UID: \"a055d801-0918-4b3b-91b9-a311d58f9d51\") " pod="openshift-marketplace/redhat-marketplace-qtn5m" Dec 06 14:11:52 crc kubenswrapper[4706]: I1206 14:11:52.877504 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a055d801-0918-4b3b-91b9-a311d58f9d51-catalog-content\") pod \"redhat-marketplace-qtn5m\" (UID: \"a055d801-0918-4b3b-91b9-a311d58f9d51\") " pod="openshift-marketplace/redhat-marketplace-qtn5m" Dec 06 14:11:52 crc kubenswrapper[4706]: I1206 14:11:52.914419 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp484\" (UniqueName: \"kubernetes.io/projected/a055d801-0918-4b3b-91b9-a311d58f9d51-kube-api-access-dp484\") pod \"redhat-marketplace-qtn5m\" (UID: \"a055d801-0918-4b3b-91b9-a311d58f9d51\") " pod="openshift-marketplace/redhat-marketplace-qtn5m" Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.016023 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qtn5m" Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.069299 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-48gk8"] Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.071091 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48gk8" Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.080694 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-48gk8"] Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.180907 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-utilities\") pod \"redhat-operators-48gk8\" (UID: \"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc\") " pod="openshift-marketplace/redhat-operators-48gk8" Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.181354 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrzb8\" (UniqueName: \"kubernetes.io/projected/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-kube-api-access-wrzb8\") pod \"redhat-operators-48gk8\" (UID: \"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc\") " pod="openshift-marketplace/redhat-operators-48gk8" Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.181432 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-catalog-content\") pod \"redhat-operators-48gk8\" (UID: \"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc\") " pod="openshift-marketplace/redhat-operators-48gk8" Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.282326 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrzb8\" (UniqueName: \"kubernetes.io/projected/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-kube-api-access-wrzb8\") pod \"redhat-operators-48gk8\" (UID: \"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc\") " pod="openshift-marketplace/redhat-operators-48gk8" Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.282434 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-catalog-content\") pod \"redhat-operators-48gk8\" (UID: \"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc\") " pod="openshift-marketplace/redhat-operators-48gk8" Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.282519 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-utilities\") pod \"redhat-operators-48gk8\" (UID: \"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc\") " pod="openshift-marketplace/redhat-operators-48gk8" Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.283164 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-catalog-content\") pod \"redhat-operators-48gk8\" (UID: \"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc\") " pod="openshift-marketplace/redhat-operators-48gk8" Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.283242 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-utilities\") pod \"redhat-operators-48gk8\" (UID: \"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc\") " pod="openshift-marketplace/redhat-operators-48gk8" Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.293944 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtn5m"] Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.307385 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrzb8\" (UniqueName: \"kubernetes.io/projected/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-kube-api-access-wrzb8\") pod \"redhat-operators-48gk8\" (UID: \"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc\") " pod="openshift-marketplace/redhat-operators-48gk8" Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.423738 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48gk8" Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.521586 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtn5m" event={"ID":"a055d801-0918-4b3b-91b9-a311d58f9d51","Type":"ContainerStarted","Data":"3236ec3d101a424172384034399e64a8d0169b9585f03b36a2cf38f22d2ff3be"} Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.521634 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtn5m" event={"ID":"a055d801-0918-4b3b-91b9-a311d58f9d51","Type":"ContainerStarted","Data":"3d0238b3b655eeb910d707cc310845407c090ceb2b65299f8d971d11bb0a40a7"} Dec 06 14:11:53 crc kubenswrapper[4706]: I1206 14:11:53.662191 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-48gk8"] Dec 06 14:11:54 crc kubenswrapper[4706]: I1206 14:11:54.528420 4706 generic.go:334] "Generic (PLEG): container finished" podID="a055d801-0918-4b3b-91b9-a311d58f9d51" containerID="3236ec3d101a424172384034399e64a8d0169b9585f03b36a2cf38f22d2ff3be" exitCode=0 Dec 06 14:11:54 crc kubenswrapper[4706]: I1206 14:11:54.528500 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtn5m" event={"ID":"a055d801-0918-4b3b-91b9-a311d58f9d51","Type":"ContainerDied","Data":"3236ec3d101a424172384034399e64a8d0169b9585f03b36a2cf38f22d2ff3be"} Dec 06 14:11:54 crc kubenswrapper[4706]: I1206 14:11:54.529809 4706 generic.go:334] "Generic (PLEG): container finished" podID="48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc" containerID="7772221a01fa0d861e22a25707cc07247e34689d01318d9e8d30dca05a66d866" exitCode=0 Dec 06 14:11:54 crc kubenswrapper[4706]: I1206 14:11:54.529832 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48gk8" event={"ID":"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc","Type":"ContainerDied","Data":"7772221a01fa0d861e22a25707cc07247e34689d01318d9e8d30dca05a66d866"} Dec 06 14:11:54 crc kubenswrapper[4706]: I1206 14:11:54.529851 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48gk8" event={"ID":"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc","Type":"ContainerStarted","Data":"3759a76b4ec842ff0f4d04acdf84e9c567e0d981d48fa2ca3f74fc6fe6fa57fe"} Dec 06 14:11:54 crc kubenswrapper[4706]: I1206 14:11:54.894584 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r"] Dec 06 14:11:54 crc kubenswrapper[4706]: I1206 14:11:54.895872 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" Dec 06 14:11:54 crc kubenswrapper[4706]: I1206 14:11:54.897552 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-fc6v2" Dec 06 14:11:54 crc kubenswrapper[4706]: I1206 14:11:54.942501 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r"] Dec 06 14:11:55 crc kubenswrapper[4706]: I1206 14:11:55.021112 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1774dcba-87ac-4217-9cde-29c8ef195974-util\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r\" (UID: \"1774dcba-87ac-4217-9cde-29c8ef195974\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" Dec 06 14:11:55 crc kubenswrapper[4706]: I1206 14:11:55.021192 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1774dcba-87ac-4217-9cde-29c8ef195974-bundle\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r\" (UID: \"1774dcba-87ac-4217-9cde-29c8ef195974\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" Dec 06 14:11:55 crc kubenswrapper[4706]: I1206 14:11:55.021367 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxnlk\" (UniqueName: \"kubernetes.io/projected/1774dcba-87ac-4217-9cde-29c8ef195974-kube-api-access-fxnlk\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r\" (UID: \"1774dcba-87ac-4217-9cde-29c8ef195974\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" Dec 06 14:11:55 crc kubenswrapper[4706]: I1206 14:11:55.122778 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1774dcba-87ac-4217-9cde-29c8ef195974-util\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r\" (UID: \"1774dcba-87ac-4217-9cde-29c8ef195974\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" Dec 06 14:11:55 crc kubenswrapper[4706]: I1206 14:11:55.122824 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1774dcba-87ac-4217-9cde-29c8ef195974-bundle\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r\" (UID: \"1774dcba-87ac-4217-9cde-29c8ef195974\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" Dec 06 14:11:55 crc kubenswrapper[4706]: I1206 14:11:55.122908 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxnlk\" (UniqueName: \"kubernetes.io/projected/1774dcba-87ac-4217-9cde-29c8ef195974-kube-api-access-fxnlk\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r\" (UID: \"1774dcba-87ac-4217-9cde-29c8ef195974\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" Dec 06 14:11:55 crc kubenswrapper[4706]: I1206 14:11:55.123712 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1774dcba-87ac-4217-9cde-29c8ef195974-util\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r\" (UID: \"1774dcba-87ac-4217-9cde-29c8ef195974\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" Dec 06 14:11:55 crc kubenswrapper[4706]: I1206 14:11:55.123766 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1774dcba-87ac-4217-9cde-29c8ef195974-bundle\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r\" (UID: \"1774dcba-87ac-4217-9cde-29c8ef195974\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" Dec 06 14:11:55 crc kubenswrapper[4706]: I1206 14:11:55.143377 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxnlk\" (UniqueName: \"kubernetes.io/projected/1774dcba-87ac-4217-9cde-29c8ef195974-kube-api-access-fxnlk\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r\" (UID: \"1774dcba-87ac-4217-9cde-29c8ef195974\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" Dec 06 14:11:55 crc kubenswrapper[4706]: I1206 14:11:55.234652 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" Dec 06 14:11:56 crc kubenswrapper[4706]: I1206 14:11:56.017726 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r"] Dec 06 14:11:56 crc kubenswrapper[4706]: I1206 14:11:56.568850 4706 generic.go:334] "Generic (PLEG): container finished" podID="a055d801-0918-4b3b-91b9-a311d58f9d51" containerID="1cad6e8049e09c0eed745ebc11f97977ca3d0f115502fee7de1d1c82afba5d70" exitCode=0 Dec 06 14:11:56 crc kubenswrapper[4706]: I1206 14:11:56.568967 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtn5m" event={"ID":"a055d801-0918-4b3b-91b9-a311d58f9d51","Type":"ContainerDied","Data":"1cad6e8049e09c0eed745ebc11f97977ca3d0f115502fee7de1d1c82afba5d70"} Dec 06 14:11:56 crc kubenswrapper[4706]: I1206 14:11:56.573808 4706 generic.go:334] "Generic (PLEG): container finished" podID="48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc" containerID="2f5856147f5e91086bf52bd1acb35163958e31a3d6e824f86dc27d8b56db81ed" exitCode=0 Dec 06 14:11:56 crc kubenswrapper[4706]: I1206 14:11:56.573843 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48gk8" event={"ID":"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc","Type":"ContainerDied","Data":"2f5856147f5e91086bf52bd1acb35163958e31a3d6e824f86dc27d8b56db81ed"} Dec 06 14:11:56 crc kubenswrapper[4706]: I1206 14:11:56.576637 4706 generic.go:334] "Generic (PLEG): container finished" podID="1774dcba-87ac-4217-9cde-29c8ef195974" containerID="3f4a4f41c411bf999dc1d8b1b7a7ba8c746d7656f80bf16ad27f247a7153ab68" exitCode=0 Dec 06 14:11:56 crc kubenswrapper[4706]: I1206 14:11:56.576674 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" event={"ID":"1774dcba-87ac-4217-9cde-29c8ef195974","Type":"ContainerDied","Data":"3f4a4f41c411bf999dc1d8b1b7a7ba8c746d7656f80bf16ad27f247a7153ab68"} Dec 06 14:11:56 crc kubenswrapper[4706]: I1206 14:11:56.576695 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" event={"ID":"1774dcba-87ac-4217-9cde-29c8ef195974","Type":"ContainerStarted","Data":"b64e1576ac5b7756e8ed43b64ef97c6f60b9e33ee23868f855dd58a075dc16b4"} Dec 06 14:11:57 crc kubenswrapper[4706]: I1206 14:11:57.583785 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48gk8" event={"ID":"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc","Type":"ContainerStarted","Data":"b8556c9a163e3488c5cbbb9e07d1a5e2ad8aa3c7beec446469527a80d5ad1ea9"} Dec 06 14:11:57 crc kubenswrapper[4706]: I1206 14:11:57.586703 4706 generic.go:334] "Generic (PLEG): container finished" podID="1774dcba-87ac-4217-9cde-29c8ef195974" containerID="445c087618e3f78c845de5174d8f25bd43aad6fe8f8d0210c0d8170406bdfd4b" exitCode=0 Dec 06 14:11:57 crc kubenswrapper[4706]: I1206 14:11:57.586800 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" event={"ID":"1774dcba-87ac-4217-9cde-29c8ef195974","Type":"ContainerDied","Data":"445c087618e3f78c845de5174d8f25bd43aad6fe8f8d0210c0d8170406bdfd4b"} Dec 06 14:11:57 crc kubenswrapper[4706]: I1206 14:11:57.594832 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtn5m" event={"ID":"a055d801-0918-4b3b-91b9-a311d58f9d51","Type":"ContainerStarted","Data":"97abe887a5ee45bf687688ffe7b48b47b2fbe78345df726eb24f7d3375bcc159"} Dec 06 14:11:57 crc kubenswrapper[4706]: I1206 14:11:57.610341 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-48gk8" podStartSLOduration=2.120566014 podStartE2EDuration="4.610324755s" podCreationTimestamp="2025-12-06 14:11:53 +0000 UTC" firstStartedPulling="2025-12-06 14:11:54.533622674 +0000 UTC m=+957.369415571" lastFinishedPulling="2025-12-06 14:11:57.023381405 +0000 UTC m=+959.859174312" observedRunningTime="2025-12-06 14:11:57.607900341 +0000 UTC m=+960.443693208" watchObservedRunningTime="2025-12-06 14:11:57.610324755 +0000 UTC m=+960.446117612" Dec 06 14:11:57 crc kubenswrapper[4706]: I1206 14:11:57.660630 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qtn5m" podStartSLOduration=3.157519794 podStartE2EDuration="5.660613688s" podCreationTimestamp="2025-12-06 14:11:52 +0000 UTC" firstStartedPulling="2025-12-06 14:11:54.533655905 +0000 UTC m=+957.369448762" lastFinishedPulling="2025-12-06 14:11:57.036749789 +0000 UTC m=+959.872542656" observedRunningTime="2025-12-06 14:11:57.635889823 +0000 UTC m=+960.471682680" watchObservedRunningTime="2025-12-06 14:11:57.660613688 +0000 UTC m=+960.496406545" Dec 06 14:11:58 crc kubenswrapper[4706]: I1206 14:11:58.604713 4706 generic.go:334] "Generic (PLEG): container finished" podID="1774dcba-87ac-4217-9cde-29c8ef195974" containerID="82de2c90b4b986b488fcde85a41a0143e73cee5e2d671a6fa9a43b3c546bb84d" exitCode=0 Dec 06 14:11:58 crc kubenswrapper[4706]: I1206 14:11:58.604787 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" event={"ID":"1774dcba-87ac-4217-9cde-29c8ef195974","Type":"ContainerDied","Data":"82de2c90b4b986b488fcde85a41a0143e73cee5e2d671a6fa9a43b3c546bb84d"} Dec 06 14:11:59 crc kubenswrapper[4706]: I1206 14:11:59.913616 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" Dec 06 14:12:00 crc kubenswrapper[4706]: I1206 14:12:00.096777 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1774dcba-87ac-4217-9cde-29c8ef195974-util\") pod \"1774dcba-87ac-4217-9cde-29c8ef195974\" (UID: \"1774dcba-87ac-4217-9cde-29c8ef195974\") " Dec 06 14:12:00 crc kubenswrapper[4706]: I1206 14:12:00.096888 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxnlk\" (UniqueName: \"kubernetes.io/projected/1774dcba-87ac-4217-9cde-29c8ef195974-kube-api-access-fxnlk\") pod \"1774dcba-87ac-4217-9cde-29c8ef195974\" (UID: \"1774dcba-87ac-4217-9cde-29c8ef195974\") " Dec 06 14:12:00 crc kubenswrapper[4706]: I1206 14:12:00.096984 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1774dcba-87ac-4217-9cde-29c8ef195974-bundle\") pod \"1774dcba-87ac-4217-9cde-29c8ef195974\" (UID: \"1774dcba-87ac-4217-9cde-29c8ef195974\") " Dec 06 14:12:00 crc kubenswrapper[4706]: I1206 14:12:00.097631 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1774dcba-87ac-4217-9cde-29c8ef195974-bundle" (OuterVolumeSpecName: "bundle") pod "1774dcba-87ac-4217-9cde-29c8ef195974" (UID: "1774dcba-87ac-4217-9cde-29c8ef195974"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:12:00 crc kubenswrapper[4706]: I1206 14:12:00.102499 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1774dcba-87ac-4217-9cde-29c8ef195974-kube-api-access-fxnlk" (OuterVolumeSpecName: "kube-api-access-fxnlk") pod "1774dcba-87ac-4217-9cde-29c8ef195974" (UID: "1774dcba-87ac-4217-9cde-29c8ef195974"). InnerVolumeSpecName "kube-api-access-fxnlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:12:00 crc kubenswrapper[4706]: I1206 14:12:00.127390 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1774dcba-87ac-4217-9cde-29c8ef195974-util" (OuterVolumeSpecName: "util") pod "1774dcba-87ac-4217-9cde-29c8ef195974" (UID: "1774dcba-87ac-4217-9cde-29c8ef195974"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:12:00 crc kubenswrapper[4706]: I1206 14:12:00.202072 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxnlk\" (UniqueName: \"kubernetes.io/projected/1774dcba-87ac-4217-9cde-29c8ef195974-kube-api-access-fxnlk\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:00 crc kubenswrapper[4706]: I1206 14:12:00.202114 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1774dcba-87ac-4217-9cde-29c8ef195974-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:00 crc kubenswrapper[4706]: I1206 14:12:00.202123 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1774dcba-87ac-4217-9cde-29c8ef195974-util\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:00 crc kubenswrapper[4706]: I1206 14:12:00.623899 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" event={"ID":"1774dcba-87ac-4217-9cde-29c8ef195974","Type":"ContainerDied","Data":"b64e1576ac5b7756e8ed43b64ef97c6f60b9e33ee23868f855dd58a075dc16b4"} Dec 06 14:12:00 crc kubenswrapper[4706]: I1206 14:12:00.624234 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b64e1576ac5b7756e8ed43b64ef97c6f60b9e33ee23868f855dd58a075dc16b4" Dec 06 14:12:00 crc kubenswrapper[4706]: I1206 14:12:00.623944 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r" Dec 06 14:12:03 crc kubenswrapper[4706]: I1206 14:12:03.016786 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qtn5m" Dec 06 14:12:03 crc kubenswrapper[4706]: I1206 14:12:03.017185 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qtn5m" Dec 06 14:12:03 crc kubenswrapper[4706]: I1206 14:12:03.093181 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qtn5m" Dec 06 14:12:03 crc kubenswrapper[4706]: I1206 14:12:03.423933 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-48gk8" Dec 06 14:12:03 crc kubenswrapper[4706]: I1206 14:12:03.424032 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-48gk8" Dec 06 14:12:03 crc kubenswrapper[4706]: I1206 14:12:03.482074 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-48gk8" Dec 06 14:12:03 crc kubenswrapper[4706]: I1206 14:12:03.714448 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qtn5m" Dec 06 14:12:03 crc kubenswrapper[4706]: I1206 14:12:03.725151 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-48gk8" Dec 06 14:12:08 crc kubenswrapper[4706]: I1206 14:12:08.051266 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtn5m"] Dec 06 14:12:08 crc kubenswrapper[4706]: I1206 14:12:08.051897 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qtn5m" podUID="a055d801-0918-4b3b-91b9-a311d58f9d51" containerName="registry-server" containerID="cri-o://97abe887a5ee45bf687688ffe7b48b47b2fbe78345df726eb24f7d3375bcc159" gracePeriod=2 Dec 06 14:12:09 crc kubenswrapper[4706]: I1206 14:12:09.053398 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-48gk8"] Dec 06 14:12:09 crc kubenswrapper[4706]: I1206 14:12:09.053770 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-48gk8" podUID="48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc" containerName="registry-server" containerID="cri-o://b8556c9a163e3488c5cbbb9e07d1a5e2ad8aa3c7beec446469527a80d5ad1ea9" gracePeriod=2 Dec 06 14:12:12 crc kubenswrapper[4706]: I1206 14:12:12.723863 4706 generic.go:334] "Generic (PLEG): container finished" podID="48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc" containerID="b8556c9a163e3488c5cbbb9e07d1a5e2ad8aa3c7beec446469527a80d5ad1ea9" exitCode=0 Dec 06 14:12:12 crc kubenswrapper[4706]: I1206 14:12:12.723952 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48gk8" event={"ID":"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc","Type":"ContainerDied","Data":"b8556c9a163e3488c5cbbb9e07d1a5e2ad8aa3c7beec446469527a80d5ad1ea9"} Dec 06 14:12:12 crc kubenswrapper[4706]: I1206 14:12:12.726698 4706 generic.go:334] "Generic (PLEG): container finished" podID="a055d801-0918-4b3b-91b9-a311d58f9d51" containerID="97abe887a5ee45bf687688ffe7b48b47b2fbe78345df726eb24f7d3375bcc159" exitCode=0 Dec 06 14:12:12 crc kubenswrapper[4706]: I1206 14:12:12.726733 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtn5m" event={"ID":"a055d801-0918-4b3b-91b9-a311d58f9d51","Type":"ContainerDied","Data":"97abe887a5ee45bf687688ffe7b48b47b2fbe78345df726eb24f7d3375bcc159"} Dec 06 14:12:12 crc kubenswrapper[4706]: I1206 14:12:12.906947 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc"] Dec 06 14:12:12 crc kubenswrapper[4706]: E1206 14:12:12.907491 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1774dcba-87ac-4217-9cde-29c8ef195974" containerName="extract" Dec 06 14:12:12 crc kubenswrapper[4706]: I1206 14:12:12.907504 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="1774dcba-87ac-4217-9cde-29c8ef195974" containerName="extract" Dec 06 14:12:12 crc kubenswrapper[4706]: E1206 14:12:12.907514 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1774dcba-87ac-4217-9cde-29c8ef195974" containerName="pull" Dec 06 14:12:12 crc kubenswrapper[4706]: I1206 14:12:12.907521 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="1774dcba-87ac-4217-9cde-29c8ef195974" containerName="pull" Dec 06 14:12:12 crc kubenswrapper[4706]: E1206 14:12:12.907535 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1774dcba-87ac-4217-9cde-29c8ef195974" containerName="util" Dec 06 14:12:12 crc kubenswrapper[4706]: I1206 14:12:12.907542 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="1774dcba-87ac-4217-9cde-29c8ef195974" containerName="util" Dec 06 14:12:12 crc kubenswrapper[4706]: I1206 14:12:12.907648 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="1774dcba-87ac-4217-9cde-29c8ef195974" containerName="extract" Dec 06 14:12:12 crc kubenswrapper[4706]: I1206 14:12:12.908049 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" Dec 06 14:12:12 crc kubenswrapper[4706]: I1206 14:12:12.909795 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Dec 06 14:12:12 crc kubenswrapper[4706]: I1206 14:12:12.909883 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-vpxjm" Dec 06 14:12:12 crc kubenswrapper[4706]: I1206 14:12:12.929192 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc"] Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.009646 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/40dc67dc-8180-452e-9345-a3e52b2adc15-webhook-cert\") pod \"keystone-operator-controller-manager-85d99cc6c-dszkc\" (UID: \"40dc67dc-8180-452e-9345-a3e52b2adc15\") " pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.009894 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8zp7\" (UniqueName: \"kubernetes.io/projected/40dc67dc-8180-452e-9345-a3e52b2adc15-kube-api-access-c8zp7\") pod \"keystone-operator-controller-manager-85d99cc6c-dszkc\" (UID: \"40dc67dc-8180-452e-9345-a3e52b2adc15\") " pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.010065 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/40dc67dc-8180-452e-9345-a3e52b2adc15-apiservice-cert\") pod \"keystone-operator-controller-manager-85d99cc6c-dszkc\" (UID: \"40dc67dc-8180-452e-9345-a3e52b2adc15\") " pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" Dec 06 14:12:13 crc kubenswrapper[4706]: E1206 14:12:13.016667 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 97abe887a5ee45bf687688ffe7b48b47b2fbe78345df726eb24f7d3375bcc159 is running failed: container process not found" containerID="97abe887a5ee45bf687688ffe7b48b47b2fbe78345df726eb24f7d3375bcc159" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 14:12:13 crc kubenswrapper[4706]: E1206 14:12:13.016986 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 97abe887a5ee45bf687688ffe7b48b47b2fbe78345df726eb24f7d3375bcc159 is running failed: container process not found" containerID="97abe887a5ee45bf687688ffe7b48b47b2fbe78345df726eb24f7d3375bcc159" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 14:12:13 crc kubenswrapper[4706]: E1206 14:12:13.017397 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 97abe887a5ee45bf687688ffe7b48b47b2fbe78345df726eb24f7d3375bcc159 is running failed: container process not found" containerID="97abe887a5ee45bf687688ffe7b48b47b2fbe78345df726eb24f7d3375bcc159" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 14:12:13 crc kubenswrapper[4706]: E1206 14:12:13.017473 4706 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 97abe887a5ee45bf687688ffe7b48b47b2fbe78345df726eb24f7d3375bcc159 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-qtn5m" podUID="a055d801-0918-4b3b-91b9-a311d58f9d51" containerName="registry-server" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.111055 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/40dc67dc-8180-452e-9345-a3e52b2adc15-apiservice-cert\") pod \"keystone-operator-controller-manager-85d99cc6c-dszkc\" (UID: \"40dc67dc-8180-452e-9345-a3e52b2adc15\") " pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.111167 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/40dc67dc-8180-452e-9345-a3e52b2adc15-webhook-cert\") pod \"keystone-operator-controller-manager-85d99cc6c-dszkc\" (UID: \"40dc67dc-8180-452e-9345-a3e52b2adc15\") " pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.111213 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8zp7\" (UniqueName: \"kubernetes.io/projected/40dc67dc-8180-452e-9345-a3e52b2adc15-kube-api-access-c8zp7\") pod \"keystone-operator-controller-manager-85d99cc6c-dszkc\" (UID: \"40dc67dc-8180-452e-9345-a3e52b2adc15\") " pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.116861 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/40dc67dc-8180-452e-9345-a3e52b2adc15-apiservice-cert\") pod \"keystone-operator-controller-manager-85d99cc6c-dszkc\" (UID: \"40dc67dc-8180-452e-9345-a3e52b2adc15\") " pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.118757 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/40dc67dc-8180-452e-9345-a3e52b2adc15-webhook-cert\") pod \"keystone-operator-controller-manager-85d99cc6c-dszkc\" (UID: \"40dc67dc-8180-452e-9345-a3e52b2adc15\") " pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.134743 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8zp7\" (UniqueName: \"kubernetes.io/projected/40dc67dc-8180-452e-9345-a3e52b2adc15-kube-api-access-c8zp7\") pod \"keystone-operator-controller-manager-85d99cc6c-dszkc\" (UID: \"40dc67dc-8180-452e-9345-a3e52b2adc15\") " pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.222982 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" Dec 06 14:12:13 crc kubenswrapper[4706]: E1206 14:12:13.426597 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b8556c9a163e3488c5cbbb9e07d1a5e2ad8aa3c7beec446469527a80d5ad1ea9 is running failed: container process not found" containerID="b8556c9a163e3488c5cbbb9e07d1a5e2ad8aa3c7beec446469527a80d5ad1ea9" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 14:12:13 crc kubenswrapper[4706]: E1206 14:12:13.431234 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b8556c9a163e3488c5cbbb9e07d1a5e2ad8aa3c7beec446469527a80d5ad1ea9 is running failed: container process not found" containerID="b8556c9a163e3488c5cbbb9e07d1a5e2ad8aa3c7beec446469527a80d5ad1ea9" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 14:12:13 crc kubenswrapper[4706]: E1206 14:12:13.431746 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b8556c9a163e3488c5cbbb9e07d1a5e2ad8aa3c7beec446469527a80d5ad1ea9 is running failed: container process not found" containerID="b8556c9a163e3488c5cbbb9e07d1a5e2ad8aa3c7beec446469527a80d5ad1ea9" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 14:12:13 crc kubenswrapper[4706]: E1206 14:12:13.431796 4706 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b8556c9a163e3488c5cbbb9e07d1a5e2ad8aa3c7beec446469527a80d5ad1ea9 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-48gk8" podUID="48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc" containerName="registry-server" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.472009 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48gk8" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.522127 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-utilities\") pod \"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc\" (UID: \"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc\") " Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.522239 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrzb8\" (UniqueName: \"kubernetes.io/projected/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-kube-api-access-wrzb8\") pod \"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc\" (UID: \"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc\") " Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.523034 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-utilities" (OuterVolumeSpecName: "utilities") pod "48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc" (UID: "48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.523110 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-catalog-content\") pod \"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc\" (UID: \"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc\") " Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.523575 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.526569 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-kube-api-access-wrzb8" (OuterVolumeSpecName: "kube-api-access-wrzb8") pod "48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc" (UID: "48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc"). InnerVolumeSpecName "kube-api-access-wrzb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.539167 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qtn5m" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.624247 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a055d801-0918-4b3b-91b9-a311d58f9d51-catalog-content\") pod \"a055d801-0918-4b3b-91b9-a311d58f9d51\" (UID: \"a055d801-0918-4b3b-91b9-a311d58f9d51\") " Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.624480 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a055d801-0918-4b3b-91b9-a311d58f9d51-utilities\") pod \"a055d801-0918-4b3b-91b9-a311d58f9d51\" (UID: \"a055d801-0918-4b3b-91b9-a311d58f9d51\") " Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.624505 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dp484\" (UniqueName: \"kubernetes.io/projected/a055d801-0918-4b3b-91b9-a311d58f9d51-kube-api-access-dp484\") pod \"a055d801-0918-4b3b-91b9-a311d58f9d51\" (UID: \"a055d801-0918-4b3b-91b9-a311d58f9d51\") " Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.624765 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrzb8\" (UniqueName: \"kubernetes.io/projected/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-kube-api-access-wrzb8\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.625536 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a055d801-0918-4b3b-91b9-a311d58f9d51-utilities" (OuterVolumeSpecName: "utilities") pod "a055d801-0918-4b3b-91b9-a311d58f9d51" (UID: "a055d801-0918-4b3b-91b9-a311d58f9d51"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.630811 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a055d801-0918-4b3b-91b9-a311d58f9d51-kube-api-access-dp484" (OuterVolumeSpecName: "kube-api-access-dp484") pod "a055d801-0918-4b3b-91b9-a311d58f9d51" (UID: "a055d801-0918-4b3b-91b9-a311d58f9d51"). InnerVolumeSpecName "kube-api-access-dp484". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.639168 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc" (UID: "48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.659268 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a055d801-0918-4b3b-91b9-a311d58f9d51-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a055d801-0918-4b3b-91b9-a311d58f9d51" (UID: "a055d801-0918-4b3b-91b9-a311d58f9d51"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.701317 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc"] Dec 06 14:12:13 crc kubenswrapper[4706]: W1206 14:12:13.705152 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40dc67dc_8180_452e_9345_a3e52b2adc15.slice/crio-cccb629ebd646bea24483a7ec1acc9633a8aba1bbd60a81df8fa6e1c724339bd WatchSource:0}: Error finding container cccb629ebd646bea24483a7ec1acc9633a8aba1bbd60a81df8fa6e1c724339bd: Status 404 returned error can't find the container with id cccb629ebd646bea24483a7ec1acc9633a8aba1bbd60a81df8fa6e1c724339bd Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.726464 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a055d801-0918-4b3b-91b9-a311d58f9d51-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.726503 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dp484\" (UniqueName: \"kubernetes.io/projected/a055d801-0918-4b3b-91b9-a311d58f9d51-kube-api-access-dp484\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.726544 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a055d801-0918-4b3b-91b9-a311d58f9d51-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.726557 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.733418 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" event={"ID":"40dc67dc-8180-452e-9345-a3e52b2adc15","Type":"ContainerStarted","Data":"cccb629ebd646bea24483a7ec1acc9633a8aba1bbd60a81df8fa6e1c724339bd"} Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.737123 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48gk8" event={"ID":"48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc","Type":"ContainerDied","Data":"3759a76b4ec842ff0f4d04acdf84e9c567e0d981d48fa2ca3f74fc6fe6fa57fe"} Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.737191 4706 scope.go:117] "RemoveContainer" containerID="b8556c9a163e3488c5cbbb9e07d1a5e2ad8aa3c7beec446469527a80d5ad1ea9" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.737250 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48gk8" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.740966 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtn5m" event={"ID":"a055d801-0918-4b3b-91b9-a311d58f9d51","Type":"ContainerDied","Data":"3d0238b3b655eeb910d707cc310845407c090ceb2b65299f8d971d11bb0a40a7"} Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.741059 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qtn5m" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.775052 4706 scope.go:117] "RemoveContainer" containerID="2f5856147f5e91086bf52bd1acb35163958e31a3d6e824f86dc27d8b56db81ed" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.776500 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-48gk8"] Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.786660 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-48gk8"] Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.795109 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtn5m"] Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.801691 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtn5m"] Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.806706 4706 scope.go:117] "RemoveContainer" containerID="7772221a01fa0d861e22a25707cc07247e34689d01318d9e8d30dca05a66d866" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.822229 4706 scope.go:117] "RemoveContainer" containerID="97abe887a5ee45bf687688ffe7b48b47b2fbe78345df726eb24f7d3375bcc159" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.837817 4706 scope.go:117] "RemoveContainer" containerID="1cad6e8049e09c0eed745ebc11f97977ca3d0f115502fee7de1d1c82afba5d70" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.857855 4706 scope.go:117] "RemoveContainer" containerID="3236ec3d101a424172384034399e64a8d0169b9585f03b36a2cf38f22d2ff3be" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.868921 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rdlzz"] Dec 06 14:12:13 crc kubenswrapper[4706]: E1206 14:12:13.869216 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a055d801-0918-4b3b-91b9-a311d58f9d51" containerName="extract-utilities" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.869236 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a055d801-0918-4b3b-91b9-a311d58f9d51" containerName="extract-utilities" Dec 06 14:12:13 crc kubenswrapper[4706]: E1206 14:12:13.869247 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc" containerName="registry-server" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.869257 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc" containerName="registry-server" Dec 06 14:12:13 crc kubenswrapper[4706]: E1206 14:12:13.869270 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc" containerName="extract-content" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.869279 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc" containerName="extract-content" Dec 06 14:12:13 crc kubenswrapper[4706]: E1206 14:12:13.869294 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a055d801-0918-4b3b-91b9-a311d58f9d51" containerName="registry-server" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.869303 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a055d801-0918-4b3b-91b9-a311d58f9d51" containerName="registry-server" Dec 06 14:12:13 crc kubenswrapper[4706]: E1206 14:12:13.869332 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc" containerName="extract-utilities" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.869341 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc" containerName="extract-utilities" Dec 06 14:12:13 crc kubenswrapper[4706]: E1206 14:12:13.869355 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a055d801-0918-4b3b-91b9-a311d58f9d51" containerName="extract-content" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.869362 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a055d801-0918-4b3b-91b9-a311d58f9d51" containerName="extract-content" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.869513 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a055d801-0918-4b3b-91b9-a311d58f9d51" containerName="registry-server" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.869537 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc" containerName="registry-server" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.870540 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rdlzz" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.887280 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rdlzz"] Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.929793 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jwql\" (UniqueName: \"kubernetes.io/projected/cce68b7a-80a8-47cc-a228-8c2749ab7565-kube-api-access-9jwql\") pod \"certified-operators-rdlzz\" (UID: \"cce68b7a-80a8-47cc-a228-8c2749ab7565\") " pod="openshift-marketplace/certified-operators-rdlzz" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.929888 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cce68b7a-80a8-47cc-a228-8c2749ab7565-utilities\") pod \"certified-operators-rdlzz\" (UID: \"cce68b7a-80a8-47cc-a228-8c2749ab7565\") " pod="openshift-marketplace/certified-operators-rdlzz" Dec 06 14:12:13 crc kubenswrapper[4706]: I1206 14:12:13.929963 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cce68b7a-80a8-47cc-a228-8c2749ab7565-catalog-content\") pod \"certified-operators-rdlzz\" (UID: \"cce68b7a-80a8-47cc-a228-8c2749ab7565\") " pod="openshift-marketplace/certified-operators-rdlzz" Dec 06 14:12:14 crc kubenswrapper[4706]: I1206 14:12:14.031030 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jwql\" (UniqueName: \"kubernetes.io/projected/cce68b7a-80a8-47cc-a228-8c2749ab7565-kube-api-access-9jwql\") pod \"certified-operators-rdlzz\" (UID: \"cce68b7a-80a8-47cc-a228-8c2749ab7565\") " pod="openshift-marketplace/certified-operators-rdlzz" Dec 06 14:12:14 crc kubenswrapper[4706]: I1206 14:12:14.031092 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cce68b7a-80a8-47cc-a228-8c2749ab7565-utilities\") pod \"certified-operators-rdlzz\" (UID: \"cce68b7a-80a8-47cc-a228-8c2749ab7565\") " pod="openshift-marketplace/certified-operators-rdlzz" Dec 06 14:12:14 crc kubenswrapper[4706]: I1206 14:12:14.031129 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cce68b7a-80a8-47cc-a228-8c2749ab7565-catalog-content\") pod \"certified-operators-rdlzz\" (UID: \"cce68b7a-80a8-47cc-a228-8c2749ab7565\") " pod="openshift-marketplace/certified-operators-rdlzz" Dec 06 14:12:14 crc kubenswrapper[4706]: I1206 14:12:14.031519 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cce68b7a-80a8-47cc-a228-8c2749ab7565-catalog-content\") pod \"certified-operators-rdlzz\" (UID: \"cce68b7a-80a8-47cc-a228-8c2749ab7565\") " pod="openshift-marketplace/certified-operators-rdlzz" Dec 06 14:12:14 crc kubenswrapper[4706]: I1206 14:12:14.031985 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cce68b7a-80a8-47cc-a228-8c2749ab7565-utilities\") pod \"certified-operators-rdlzz\" (UID: \"cce68b7a-80a8-47cc-a228-8c2749ab7565\") " pod="openshift-marketplace/certified-operators-rdlzz" Dec 06 14:12:14 crc kubenswrapper[4706]: I1206 14:12:14.049683 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jwql\" (UniqueName: \"kubernetes.io/projected/cce68b7a-80a8-47cc-a228-8c2749ab7565-kube-api-access-9jwql\") pod \"certified-operators-rdlzz\" (UID: \"cce68b7a-80a8-47cc-a228-8c2749ab7565\") " pod="openshift-marketplace/certified-operators-rdlzz" Dec 06 14:12:14 crc kubenswrapper[4706]: I1206 14:12:14.188165 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rdlzz" Dec 06 14:12:14 crc kubenswrapper[4706]: I1206 14:12:14.418212 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rdlzz"] Dec 06 14:12:14 crc kubenswrapper[4706]: I1206 14:12:14.580505 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:12:14 crc kubenswrapper[4706]: I1206 14:12:14.580841 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:12:14 crc kubenswrapper[4706]: I1206 14:12:14.756294 4706 generic.go:334] "Generic (PLEG): container finished" podID="cce68b7a-80a8-47cc-a228-8c2749ab7565" containerID="4696b5e84a8bf5969ca6d0419fb03bbb926e61501dbb6235df4a99f33099a24a" exitCode=0 Dec 06 14:12:14 crc kubenswrapper[4706]: I1206 14:12:14.756380 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdlzz" event={"ID":"cce68b7a-80a8-47cc-a228-8c2749ab7565","Type":"ContainerDied","Data":"4696b5e84a8bf5969ca6d0419fb03bbb926e61501dbb6235df4a99f33099a24a"} Dec 06 14:12:14 crc kubenswrapper[4706]: I1206 14:12:14.756407 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdlzz" event={"ID":"cce68b7a-80a8-47cc-a228-8c2749ab7565","Type":"ContainerStarted","Data":"2347b7b214703ef62c995e960047cede4d3d86692ca4d2e7846b4347e6e13754"} Dec 06 14:12:15 crc kubenswrapper[4706]: I1206 14:12:15.521898 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc" path="/var/lib/kubelet/pods/48fd2b2e-7aa6-47b2-8a28-ebdaf3166cfc/volumes" Dec 06 14:12:15 crc kubenswrapper[4706]: I1206 14:12:15.522977 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a055d801-0918-4b3b-91b9-a311d58f9d51" path="/var/lib/kubelet/pods/a055d801-0918-4b3b-91b9-a311d58f9d51/volumes" Dec 06 14:12:16 crc kubenswrapper[4706]: I1206 14:12:16.770158 4706 generic.go:334] "Generic (PLEG): container finished" podID="23f8c907-7a56-4da3-aefb-929a43497eb2" containerID="6a2c055976a1f626f8c484f7e2eab83c12e64177e0a91754b946043503eb7531" exitCode=0 Dec 06 14:12:16 crc kubenswrapper[4706]: I1206 14:12:16.770274 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"23f8c907-7a56-4da3-aefb-929a43497eb2","Type":"ContainerDied","Data":"6a2c055976a1f626f8c484f7e2eab83c12e64177e0a91754b946043503eb7531"} Dec 06 14:12:17 crc kubenswrapper[4706]: I1206 14:12:17.777766 4706 generic.go:334] "Generic (PLEG): container finished" podID="cce68b7a-80a8-47cc-a228-8c2749ab7565" containerID="2d62781aa8d01ea9e9ddb302f4d9cc60cfaae27a77a8e4ce609afbc4816eb682" exitCode=0 Dec 06 14:12:17 crc kubenswrapper[4706]: I1206 14:12:17.779243 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdlzz" event={"ID":"cce68b7a-80a8-47cc-a228-8c2749ab7565","Type":"ContainerDied","Data":"2d62781aa8d01ea9e9ddb302f4d9cc60cfaae27a77a8e4ce609afbc4816eb682"} Dec 06 14:12:17 crc kubenswrapper[4706]: I1206 14:12:17.780516 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" event={"ID":"40dc67dc-8180-452e-9345-a3e52b2adc15","Type":"ContainerStarted","Data":"9cda45ed2d5c17ab5baaa830233df7136e195395cffb205681faba4a417e4a39"} Dec 06 14:12:17 crc kubenswrapper[4706]: I1206 14:12:17.780909 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" Dec 06 14:12:17 crc kubenswrapper[4706]: I1206 14:12:17.782936 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"23f8c907-7a56-4da3-aefb-929a43497eb2","Type":"ContainerStarted","Data":"d659020ec94512a5441f62fd7acc50ff9f16952d2c086bd354c4ef2ed27c7465"} Dec 06 14:12:17 crc kubenswrapper[4706]: I1206 14:12:17.783137 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:12:17 crc kubenswrapper[4706]: I1206 14:12:17.833429 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" podStartSLOduration=2.461935853 podStartE2EDuration="5.833414467s" podCreationTimestamp="2025-12-06 14:12:12 +0000 UTC" firstStartedPulling="2025-12-06 14:12:13.707933962 +0000 UTC m=+976.543726819" lastFinishedPulling="2025-12-06 14:12:17.079412566 +0000 UTC m=+979.915205433" observedRunningTime="2025-12-06 14:12:17.829572305 +0000 UTC m=+980.665365162" watchObservedRunningTime="2025-12-06 14:12:17.833414467 +0000 UTC m=+980.669207324" Dec 06 14:12:17 crc kubenswrapper[4706]: I1206 14:12:17.853832 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/rabbitmq-server-0" podStartSLOduration=37.022815065 podStartE2EDuration="45.853808907s" podCreationTimestamp="2025-12-06 14:11:32 +0000 UTC" firstStartedPulling="2025-12-06 14:11:33.837451364 +0000 UTC m=+936.673244231" lastFinishedPulling="2025-12-06 14:11:42.668445196 +0000 UTC m=+945.504238073" observedRunningTime="2025-12-06 14:12:17.84714139 +0000 UTC m=+980.682934267" watchObservedRunningTime="2025-12-06 14:12:17.853808907 +0000 UTC m=+980.689601764" Dec 06 14:12:18 crc kubenswrapper[4706]: I1206 14:12:18.794499 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdlzz" event={"ID":"cce68b7a-80a8-47cc-a228-8c2749ab7565","Type":"ContainerStarted","Data":"ec60c93ecbb068e4f24683927123be1b6a0670a5b40774b086779f039052c33f"} Dec 06 14:12:18 crc kubenswrapper[4706]: I1206 14:12:18.828339 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rdlzz" podStartSLOduration=2.455545727 podStartE2EDuration="5.828302755s" podCreationTimestamp="2025-12-06 14:12:13 +0000 UTC" firstStartedPulling="2025-12-06 14:12:14.757892988 +0000 UTC m=+977.593685865" lastFinishedPulling="2025-12-06 14:12:18.130650036 +0000 UTC m=+980.966442893" observedRunningTime="2025-12-06 14:12:18.822231904 +0000 UTC m=+981.658024781" watchObservedRunningTime="2025-12-06 14:12:18.828302755 +0000 UTC m=+981.664095622" Dec 06 14:12:23 crc kubenswrapper[4706]: I1206 14:12:23.227047 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" Dec 06 14:12:24 crc kubenswrapper[4706]: I1206 14:12:24.192019 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rdlzz" Dec 06 14:12:24 crc kubenswrapper[4706]: I1206 14:12:24.192117 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rdlzz" Dec 06 14:12:24 crc kubenswrapper[4706]: I1206 14:12:24.272402 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rdlzz" Dec 06 14:12:24 crc kubenswrapper[4706]: I1206 14:12:24.884747 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rdlzz" Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.541821 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-db-create-57889"] Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.543062 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-57889" Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.549703 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf"] Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.550936 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf" Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.553741 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-db-secret" Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.553775 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-57889"] Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.562638 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf"] Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.722433 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0-operator-scripts\") pod \"keystone-ec25-account-create-update-7b5jf\" (UID: \"7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0\") " pod="cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf" Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.722770 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvzkk\" (UniqueName: \"kubernetes.io/projected/40779db0-e757-45d1-956f-bc17fb6b2c62-kube-api-access-bvzkk\") pod \"keystone-db-create-57889\" (UID: \"40779db0-e757-45d1-956f-bc17fb6b2c62\") " pod="cinder-kuttl-tests/keystone-db-create-57889" Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.722973 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40779db0-e757-45d1-956f-bc17fb6b2c62-operator-scripts\") pod \"keystone-db-create-57889\" (UID: \"40779db0-e757-45d1-956f-bc17fb6b2c62\") " pod="cinder-kuttl-tests/keystone-db-create-57889" Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.723043 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw7z5\" (UniqueName: \"kubernetes.io/projected/7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0-kube-api-access-dw7z5\") pod \"keystone-ec25-account-create-update-7b5jf\" (UID: \"7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0\") " pod="cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf" Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.824553 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvzkk\" (UniqueName: \"kubernetes.io/projected/40779db0-e757-45d1-956f-bc17fb6b2c62-kube-api-access-bvzkk\") pod \"keystone-db-create-57889\" (UID: \"40779db0-e757-45d1-956f-bc17fb6b2c62\") " pod="cinder-kuttl-tests/keystone-db-create-57889" Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.825246 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40779db0-e757-45d1-956f-bc17fb6b2c62-operator-scripts\") pod \"keystone-db-create-57889\" (UID: \"40779db0-e757-45d1-956f-bc17fb6b2c62\") " pod="cinder-kuttl-tests/keystone-db-create-57889" Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.826622 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40779db0-e757-45d1-956f-bc17fb6b2c62-operator-scripts\") pod \"keystone-db-create-57889\" (UID: \"40779db0-e757-45d1-956f-bc17fb6b2c62\") " pod="cinder-kuttl-tests/keystone-db-create-57889" Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.826853 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw7z5\" (UniqueName: \"kubernetes.io/projected/7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0-kube-api-access-dw7z5\") pod \"keystone-ec25-account-create-update-7b5jf\" (UID: \"7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0\") " pod="cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf" Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.827103 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0-operator-scripts\") pod \"keystone-ec25-account-create-update-7b5jf\" (UID: \"7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0\") " pod="cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf" Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.828190 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0-operator-scripts\") pod \"keystone-ec25-account-create-update-7b5jf\" (UID: \"7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0\") " pod="cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf" Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.853591 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw7z5\" (UniqueName: \"kubernetes.io/projected/7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0-kube-api-access-dw7z5\") pod \"keystone-ec25-account-create-update-7b5jf\" (UID: \"7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0\") " pod="cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf" Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.861276 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvzkk\" (UniqueName: \"kubernetes.io/projected/40779db0-e757-45d1-956f-bc17fb6b2c62-kube-api-access-bvzkk\") pod \"keystone-db-create-57889\" (UID: \"40779db0-e757-45d1-956f-bc17fb6b2c62\") " pod="cinder-kuttl-tests/keystone-db-create-57889" Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.861639 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-57889" Dec 06 14:12:27 crc kubenswrapper[4706]: I1206 14:12:27.870639 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf" Dec 06 14:12:28 crc kubenswrapper[4706]: I1206 14:12:28.203849 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-57889"] Dec 06 14:12:28 crc kubenswrapper[4706]: W1206 14:12:28.209095 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40779db0_e757_45d1_956f_bc17fb6b2c62.slice/crio-194c4d071c42c314e522ccd24627ce75b0ede53dc89f920c6e8746662b11408f WatchSource:0}: Error finding container 194c4d071c42c314e522ccd24627ce75b0ede53dc89f920c6e8746662b11408f: Status 404 returned error can't find the container with id 194c4d071c42c314e522ccd24627ce75b0ede53dc89f920c6e8746662b11408f Dec 06 14:12:28 crc kubenswrapper[4706]: I1206 14:12:28.354296 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf"] Dec 06 14:12:28 crc kubenswrapper[4706]: W1206 14:12:28.358415 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ac35e8f_d1fe_472b_8f1d_9d0424fa6ed0.slice/crio-63f6d2c7e0f5372f885dafd186850211a3701d98133a3f116ccfc6ba4184a90b WatchSource:0}: Error finding container 63f6d2c7e0f5372f885dafd186850211a3701d98133a3f116ccfc6ba4184a90b: Status 404 returned error can't find the container with id 63f6d2c7e0f5372f885dafd186850211a3701d98133a3f116ccfc6ba4184a90b Dec 06 14:12:28 crc kubenswrapper[4706]: I1206 14:12:28.877685 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-57889" event={"ID":"40779db0-e757-45d1-956f-bc17fb6b2c62","Type":"ContainerStarted","Data":"e63b9aa8c618d24417a97f3b26af0df3138dd89c00e4ea8971234646322f6402"} Dec 06 14:12:28 crc kubenswrapper[4706]: I1206 14:12:28.877734 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-57889" event={"ID":"40779db0-e757-45d1-956f-bc17fb6b2c62","Type":"ContainerStarted","Data":"194c4d071c42c314e522ccd24627ce75b0ede53dc89f920c6e8746662b11408f"} Dec 06 14:12:28 crc kubenswrapper[4706]: I1206 14:12:28.879625 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf" event={"ID":"7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0","Type":"ContainerStarted","Data":"f9aa6c96348f40276b585ba07998ba23d7f79cde69151709d1958250cb56696b"} Dec 06 14:12:28 crc kubenswrapper[4706]: I1206 14:12:28.879647 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf" event={"ID":"7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0","Type":"ContainerStarted","Data":"63f6d2c7e0f5372f885dafd186850211a3701d98133a3f116ccfc6ba4184a90b"} Dec 06 14:12:28 crc kubenswrapper[4706]: I1206 14:12:28.889235 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-db-create-57889" podStartSLOduration=1.889220138 podStartE2EDuration="1.889220138s" podCreationTimestamp="2025-12-06 14:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:12:28.887118683 +0000 UTC m=+991.722911540" watchObservedRunningTime="2025-12-06 14:12:28.889220138 +0000 UTC m=+991.725012995" Dec 06 14:12:28 crc kubenswrapper[4706]: I1206 14:12:28.902429 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf" podStartSLOduration=1.902414348 podStartE2EDuration="1.902414348s" podCreationTimestamp="2025-12-06 14:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:12:28.897053036 +0000 UTC m=+991.732845893" watchObservedRunningTime="2025-12-06 14:12:28.902414348 +0000 UTC m=+991.738207205" Dec 06 14:12:29 crc kubenswrapper[4706]: I1206 14:12:29.889879 4706 generic.go:334] "Generic (PLEG): container finished" podID="7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0" containerID="f9aa6c96348f40276b585ba07998ba23d7f79cde69151709d1958250cb56696b" exitCode=0 Dec 06 14:12:29 crc kubenswrapper[4706]: I1206 14:12:29.890048 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf" event={"ID":"7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0","Type":"ContainerDied","Data":"f9aa6c96348f40276b585ba07998ba23d7f79cde69151709d1958250cb56696b"} Dec 06 14:12:29 crc kubenswrapper[4706]: I1206 14:12:29.891737 4706 generic.go:334] "Generic (PLEG): container finished" podID="40779db0-e757-45d1-956f-bc17fb6b2c62" containerID="e63b9aa8c618d24417a97f3b26af0df3138dd89c00e4ea8971234646322f6402" exitCode=0 Dec 06 14:12:29 crc kubenswrapper[4706]: I1206 14:12:29.891776 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-57889" event={"ID":"40779db0-e757-45d1-956f-bc17fb6b2c62","Type":"ContainerDied","Data":"e63b9aa8c618d24417a97f3b26af0df3138dd89c00e4ea8971234646322f6402"} Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.217389 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.249671 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rdlzz"] Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.249897 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rdlzz" podUID="cce68b7a-80a8-47cc-a228-8c2749ab7565" containerName="registry-server" containerID="cri-o://ec60c93ecbb068e4f24683927123be1b6a0670a5b40774b086779f039052c33f" gracePeriod=2 Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.379954 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0-operator-scripts\") pod \"7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0\" (UID: \"7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0\") " Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.380000 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw7z5\" (UniqueName: \"kubernetes.io/projected/7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0-kube-api-access-dw7z5\") pod \"7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0\" (UID: \"7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0\") " Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.380444 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0" (UID: "7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.386658 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0-kube-api-access-dw7z5" (OuterVolumeSpecName: "kube-api-access-dw7z5") pod "7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0" (UID: "7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0"). InnerVolumeSpecName "kube-api-access-dw7z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.389278 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-57889" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.481249 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.481288 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw7z5\" (UniqueName: \"kubernetes.io/projected/7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0-kube-api-access-dw7z5\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.582693 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvzkk\" (UniqueName: \"kubernetes.io/projected/40779db0-e757-45d1-956f-bc17fb6b2c62-kube-api-access-bvzkk\") pod \"40779db0-e757-45d1-956f-bc17fb6b2c62\" (UID: \"40779db0-e757-45d1-956f-bc17fb6b2c62\") " Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.582770 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40779db0-e757-45d1-956f-bc17fb6b2c62-operator-scripts\") pod \"40779db0-e757-45d1-956f-bc17fb6b2c62\" (UID: \"40779db0-e757-45d1-956f-bc17fb6b2c62\") " Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.583827 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40779db0-e757-45d1-956f-bc17fb6b2c62-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "40779db0-e757-45d1-956f-bc17fb6b2c62" (UID: "40779db0-e757-45d1-956f-bc17fb6b2c62"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.589047 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40779db0-e757-45d1-956f-bc17fb6b2c62-kube-api-access-bvzkk" (OuterVolumeSpecName: "kube-api-access-bvzkk") pod "40779db0-e757-45d1-956f-bc17fb6b2c62" (UID: "40779db0-e757-45d1-956f-bc17fb6b2c62"). InnerVolumeSpecName "kube-api-access-bvzkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.594349 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rdlzz" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.684166 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvzkk\" (UniqueName: \"kubernetes.io/projected/40779db0-e757-45d1-956f-bc17fb6b2c62-kube-api-access-bvzkk\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.684205 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40779db0-e757-45d1-956f-bc17fb6b2c62-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.785484 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cce68b7a-80a8-47cc-a228-8c2749ab7565-utilities\") pod \"cce68b7a-80a8-47cc-a228-8c2749ab7565\" (UID: \"cce68b7a-80a8-47cc-a228-8c2749ab7565\") " Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.785592 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jwql\" (UniqueName: \"kubernetes.io/projected/cce68b7a-80a8-47cc-a228-8c2749ab7565-kube-api-access-9jwql\") pod \"cce68b7a-80a8-47cc-a228-8c2749ab7565\" (UID: \"cce68b7a-80a8-47cc-a228-8c2749ab7565\") " Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.785702 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cce68b7a-80a8-47cc-a228-8c2749ab7565-catalog-content\") pod \"cce68b7a-80a8-47cc-a228-8c2749ab7565\" (UID: \"cce68b7a-80a8-47cc-a228-8c2749ab7565\") " Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.789263 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cce68b7a-80a8-47cc-a228-8c2749ab7565-utilities" (OuterVolumeSpecName: "utilities") pod "cce68b7a-80a8-47cc-a228-8c2749ab7565" (UID: "cce68b7a-80a8-47cc-a228-8c2749ab7565"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.790393 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cce68b7a-80a8-47cc-a228-8c2749ab7565-kube-api-access-9jwql" (OuterVolumeSpecName: "kube-api-access-9jwql") pod "cce68b7a-80a8-47cc-a228-8c2749ab7565" (UID: "cce68b7a-80a8-47cc-a228-8c2749ab7565"). InnerVolumeSpecName "kube-api-access-9jwql". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.852823 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cce68b7a-80a8-47cc-a228-8c2749ab7565-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cce68b7a-80a8-47cc-a228-8c2749ab7565" (UID: "cce68b7a-80a8-47cc-a228-8c2749ab7565"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.887699 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cce68b7a-80a8-47cc-a228-8c2749ab7565-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.887729 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jwql\" (UniqueName: \"kubernetes.io/projected/cce68b7a-80a8-47cc-a228-8c2749ab7565-kube-api-access-9jwql\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.887739 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cce68b7a-80a8-47cc-a228-8c2749ab7565-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.907374 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-57889" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.907389 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-57889" event={"ID":"40779db0-e757-45d1-956f-bc17fb6b2c62","Type":"ContainerDied","Data":"194c4d071c42c314e522ccd24627ce75b0ede53dc89f920c6e8746662b11408f"} Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.907447 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="194c4d071c42c314e522ccd24627ce75b0ede53dc89f920c6e8746662b11408f" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.909367 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf" event={"ID":"7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0","Type":"ContainerDied","Data":"63f6d2c7e0f5372f885dafd186850211a3701d98133a3f116ccfc6ba4184a90b"} Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.909423 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63f6d2c7e0f5372f885dafd186850211a3701d98133a3f116ccfc6ba4184a90b" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.909515 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.913529 4706 generic.go:334] "Generic (PLEG): container finished" podID="cce68b7a-80a8-47cc-a228-8c2749ab7565" containerID="ec60c93ecbb068e4f24683927123be1b6a0670a5b40774b086779f039052c33f" exitCode=0 Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.913570 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdlzz" event={"ID":"cce68b7a-80a8-47cc-a228-8c2749ab7565","Type":"ContainerDied","Data":"ec60c93ecbb068e4f24683927123be1b6a0670a5b40774b086779f039052c33f"} Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.913594 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdlzz" event={"ID":"cce68b7a-80a8-47cc-a228-8c2749ab7565","Type":"ContainerDied","Data":"2347b7b214703ef62c995e960047cede4d3d86692ca4d2e7846b4347e6e13754"} Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.913592 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rdlzz" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.913668 4706 scope.go:117] "RemoveContainer" containerID="ec60c93ecbb068e4f24683927123be1b6a0670a5b40774b086779f039052c33f" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.942750 4706 scope.go:117] "RemoveContainer" containerID="2d62781aa8d01ea9e9ddb302f4d9cc60cfaae27a77a8e4ce609afbc4816eb682" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.957832 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rdlzz"] Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.961981 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rdlzz"] Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.969060 4706 scope.go:117] "RemoveContainer" containerID="4696b5e84a8bf5969ca6d0419fb03bbb926e61501dbb6235df4a99f33099a24a" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.986650 4706 scope.go:117] "RemoveContainer" containerID="ec60c93ecbb068e4f24683927123be1b6a0670a5b40774b086779f039052c33f" Dec 06 14:12:31 crc kubenswrapper[4706]: E1206 14:12:31.987013 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec60c93ecbb068e4f24683927123be1b6a0670a5b40774b086779f039052c33f\": container with ID starting with ec60c93ecbb068e4f24683927123be1b6a0670a5b40774b086779f039052c33f not found: ID does not exist" containerID="ec60c93ecbb068e4f24683927123be1b6a0670a5b40774b086779f039052c33f" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.987172 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec60c93ecbb068e4f24683927123be1b6a0670a5b40774b086779f039052c33f"} err="failed to get container status \"ec60c93ecbb068e4f24683927123be1b6a0670a5b40774b086779f039052c33f\": rpc error: code = NotFound desc = could not find container \"ec60c93ecbb068e4f24683927123be1b6a0670a5b40774b086779f039052c33f\": container with ID starting with ec60c93ecbb068e4f24683927123be1b6a0670a5b40774b086779f039052c33f not found: ID does not exist" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.987288 4706 scope.go:117] "RemoveContainer" containerID="2d62781aa8d01ea9e9ddb302f4d9cc60cfaae27a77a8e4ce609afbc4816eb682" Dec 06 14:12:31 crc kubenswrapper[4706]: E1206 14:12:31.987799 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d62781aa8d01ea9e9ddb302f4d9cc60cfaae27a77a8e4ce609afbc4816eb682\": container with ID starting with 2d62781aa8d01ea9e9ddb302f4d9cc60cfaae27a77a8e4ce609afbc4816eb682 not found: ID does not exist" containerID="2d62781aa8d01ea9e9ddb302f4d9cc60cfaae27a77a8e4ce609afbc4816eb682" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.987857 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d62781aa8d01ea9e9ddb302f4d9cc60cfaae27a77a8e4ce609afbc4816eb682"} err="failed to get container status \"2d62781aa8d01ea9e9ddb302f4d9cc60cfaae27a77a8e4ce609afbc4816eb682\": rpc error: code = NotFound desc = could not find container \"2d62781aa8d01ea9e9ddb302f4d9cc60cfaae27a77a8e4ce609afbc4816eb682\": container with ID starting with 2d62781aa8d01ea9e9ddb302f4d9cc60cfaae27a77a8e4ce609afbc4816eb682 not found: ID does not exist" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.987893 4706 scope.go:117] "RemoveContainer" containerID="4696b5e84a8bf5969ca6d0419fb03bbb926e61501dbb6235df4a99f33099a24a" Dec 06 14:12:31 crc kubenswrapper[4706]: E1206 14:12:31.988213 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4696b5e84a8bf5969ca6d0419fb03bbb926e61501dbb6235df4a99f33099a24a\": container with ID starting with 4696b5e84a8bf5969ca6d0419fb03bbb926e61501dbb6235df4a99f33099a24a not found: ID does not exist" containerID="4696b5e84a8bf5969ca6d0419fb03bbb926e61501dbb6235df4a99f33099a24a" Dec 06 14:12:31 crc kubenswrapper[4706]: I1206 14:12:31.988343 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4696b5e84a8bf5969ca6d0419fb03bbb926e61501dbb6235df4a99f33099a24a"} err="failed to get container status \"4696b5e84a8bf5969ca6d0419fb03bbb926e61501dbb6235df4a99f33099a24a\": rpc error: code = NotFound desc = could not find container \"4696b5e84a8bf5969ca6d0419fb03bbb926e61501dbb6235df4a99f33099a24a\": container with ID starting with 4696b5e84a8bf5969ca6d0419fb03bbb926e61501dbb6235df4a99f33099a24a not found: ID does not exist" Dec 06 14:12:32 crc kubenswrapper[4706]: I1206 14:12:32.261160 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-index-klwkp"] Dec 06 14:12:32 crc kubenswrapper[4706]: E1206 14:12:32.261524 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cce68b7a-80a8-47cc-a228-8c2749ab7565" containerName="extract-content" Dec 06 14:12:32 crc kubenswrapper[4706]: I1206 14:12:32.261543 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="cce68b7a-80a8-47cc-a228-8c2749ab7565" containerName="extract-content" Dec 06 14:12:32 crc kubenswrapper[4706]: E1206 14:12:32.261565 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cce68b7a-80a8-47cc-a228-8c2749ab7565" containerName="extract-utilities" Dec 06 14:12:32 crc kubenswrapper[4706]: I1206 14:12:32.261577 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="cce68b7a-80a8-47cc-a228-8c2749ab7565" containerName="extract-utilities" Dec 06 14:12:32 crc kubenswrapper[4706]: E1206 14:12:32.261588 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cce68b7a-80a8-47cc-a228-8c2749ab7565" containerName="registry-server" Dec 06 14:12:32 crc kubenswrapper[4706]: I1206 14:12:32.261599 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="cce68b7a-80a8-47cc-a228-8c2749ab7565" containerName="registry-server" Dec 06 14:12:32 crc kubenswrapper[4706]: E1206 14:12:32.261613 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0" containerName="mariadb-account-create-update" Dec 06 14:12:32 crc kubenswrapper[4706]: I1206 14:12:32.261625 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0" containerName="mariadb-account-create-update" Dec 06 14:12:32 crc kubenswrapper[4706]: E1206 14:12:32.261652 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40779db0-e757-45d1-956f-bc17fb6b2c62" containerName="mariadb-database-create" Dec 06 14:12:32 crc kubenswrapper[4706]: I1206 14:12:32.261663 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="40779db0-e757-45d1-956f-bc17fb6b2c62" containerName="mariadb-database-create" Dec 06 14:12:32 crc kubenswrapper[4706]: I1206 14:12:32.261841 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="40779db0-e757-45d1-956f-bc17fb6b2c62" containerName="mariadb-database-create" Dec 06 14:12:32 crc kubenswrapper[4706]: I1206 14:12:32.261864 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="cce68b7a-80a8-47cc-a228-8c2749ab7565" containerName="registry-server" Dec 06 14:12:32 crc kubenswrapper[4706]: I1206 14:12:32.261884 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0" containerName="mariadb-account-create-update" Dec 06 14:12:32 crc kubenswrapper[4706]: I1206 14:12:32.262536 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-klwkp" Dec 06 14:12:32 crc kubenswrapper[4706]: I1206 14:12:32.265024 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-index-dockercfg-fwq9c" Dec 06 14:12:32 crc kubenswrapper[4706]: I1206 14:12:32.280279 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-index-klwkp"] Dec 06 14:12:32 crc kubenswrapper[4706]: I1206 14:12:32.398493 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt9vs\" (UniqueName: \"kubernetes.io/projected/c36a0ce1-c1a6-4529-9579-a48508a688fc-kube-api-access-zt9vs\") pod \"cinder-operator-index-klwkp\" (UID: \"c36a0ce1-c1a6-4529-9579-a48508a688fc\") " pod="openstack-operators/cinder-operator-index-klwkp" Dec 06 14:12:32 crc kubenswrapper[4706]: I1206 14:12:32.500516 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt9vs\" (UniqueName: \"kubernetes.io/projected/c36a0ce1-c1a6-4529-9579-a48508a688fc-kube-api-access-zt9vs\") pod \"cinder-operator-index-klwkp\" (UID: \"c36a0ce1-c1a6-4529-9579-a48508a688fc\") " pod="openstack-operators/cinder-operator-index-klwkp" Dec 06 14:12:32 crc kubenswrapper[4706]: I1206 14:12:32.521617 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt9vs\" (UniqueName: \"kubernetes.io/projected/c36a0ce1-c1a6-4529-9579-a48508a688fc-kube-api-access-zt9vs\") pod \"cinder-operator-index-klwkp\" (UID: \"c36a0ce1-c1a6-4529-9579-a48508a688fc\") " pod="openstack-operators/cinder-operator-index-klwkp" Dec 06 14:12:32 crc kubenswrapper[4706]: I1206 14:12:32.601240 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-klwkp" Dec 06 14:12:33 crc kubenswrapper[4706]: I1206 14:12:33.085796 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-index-klwkp"] Dec 06 14:12:33 crc kubenswrapper[4706]: W1206 14:12:33.089993 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc36a0ce1_c1a6_4529_9579_a48508a688fc.slice/crio-c9aa9b2e3f6b74c0d3f67dd38ffdc10e89e7f098e277eceb3edba07a98854646 WatchSource:0}: Error finding container c9aa9b2e3f6b74c0d3f67dd38ffdc10e89e7f098e277eceb3edba07a98854646: Status 404 returned error can't find the container with id c9aa9b2e3f6b74c0d3f67dd38ffdc10e89e7f098e277eceb3edba07a98854646 Dec 06 14:12:33 crc kubenswrapper[4706]: I1206 14:12:33.524779 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cce68b7a-80a8-47cc-a228-8c2749ab7565" path="/var/lib/kubelet/pods/cce68b7a-80a8-47cc-a228-8c2749ab7565/volumes" Dec 06 14:12:33 crc kubenswrapper[4706]: I1206 14:12:33.553588 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:12:33 crc kubenswrapper[4706]: I1206 14:12:33.933750 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-klwkp" event={"ID":"c36a0ce1-c1a6-4529-9579-a48508a688fc","Type":"ContainerStarted","Data":"c9aa9b2e3f6b74c0d3f67dd38ffdc10e89e7f098e277eceb3edba07a98854646"} Dec 06 14:12:34 crc kubenswrapper[4706]: I1206 14:12:34.102110 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-fzrvx"] Dec 06 14:12:34 crc kubenswrapper[4706]: I1206 14:12:34.102856 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-fzrvx" Dec 06 14:12:34 crc kubenswrapper[4706]: I1206 14:12:34.104825 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone" Dec 06 14:12:34 crc kubenswrapper[4706]: I1206 14:12:34.104825 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-keystone-dockercfg-5brnv" Dec 06 14:12:34 crc kubenswrapper[4706]: I1206 14:12:34.105227 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-config-data" Dec 06 14:12:34 crc kubenswrapper[4706]: I1206 14:12:34.106346 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-scripts" Dec 06 14:12:34 crc kubenswrapper[4706]: I1206 14:12:34.122119 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-fzrvx"] Dec 06 14:12:34 crc kubenswrapper[4706]: I1206 14:12:34.229633 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7af82ce5-6c72-4654-b677-f517f8885495-config-data\") pod \"keystone-db-sync-fzrvx\" (UID: \"7af82ce5-6c72-4654-b677-f517f8885495\") " pod="cinder-kuttl-tests/keystone-db-sync-fzrvx" Dec 06 14:12:34 crc kubenswrapper[4706]: I1206 14:12:34.229734 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vvbg\" (UniqueName: \"kubernetes.io/projected/7af82ce5-6c72-4654-b677-f517f8885495-kube-api-access-8vvbg\") pod \"keystone-db-sync-fzrvx\" (UID: \"7af82ce5-6c72-4654-b677-f517f8885495\") " pod="cinder-kuttl-tests/keystone-db-sync-fzrvx" Dec 06 14:12:34 crc kubenswrapper[4706]: I1206 14:12:34.331191 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vvbg\" (UniqueName: \"kubernetes.io/projected/7af82ce5-6c72-4654-b677-f517f8885495-kube-api-access-8vvbg\") pod \"keystone-db-sync-fzrvx\" (UID: \"7af82ce5-6c72-4654-b677-f517f8885495\") " pod="cinder-kuttl-tests/keystone-db-sync-fzrvx" Dec 06 14:12:34 crc kubenswrapper[4706]: I1206 14:12:34.331255 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7af82ce5-6c72-4654-b677-f517f8885495-config-data\") pod \"keystone-db-sync-fzrvx\" (UID: \"7af82ce5-6c72-4654-b677-f517f8885495\") " pod="cinder-kuttl-tests/keystone-db-sync-fzrvx" Dec 06 14:12:34 crc kubenswrapper[4706]: I1206 14:12:34.337573 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7af82ce5-6c72-4654-b677-f517f8885495-config-data\") pod \"keystone-db-sync-fzrvx\" (UID: \"7af82ce5-6c72-4654-b677-f517f8885495\") " pod="cinder-kuttl-tests/keystone-db-sync-fzrvx" Dec 06 14:12:34 crc kubenswrapper[4706]: I1206 14:12:34.347721 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vvbg\" (UniqueName: \"kubernetes.io/projected/7af82ce5-6c72-4654-b677-f517f8885495-kube-api-access-8vvbg\") pod \"keystone-db-sync-fzrvx\" (UID: \"7af82ce5-6c72-4654-b677-f517f8885495\") " pod="cinder-kuttl-tests/keystone-db-sync-fzrvx" Dec 06 14:12:34 crc kubenswrapper[4706]: I1206 14:12:34.422158 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-fzrvx" Dec 06 14:12:34 crc kubenswrapper[4706]: I1206 14:12:34.918935 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-fzrvx"] Dec 06 14:12:35 crc kubenswrapper[4706]: W1206 14:12:35.062052 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7af82ce5_6c72_4654_b677_f517f8885495.slice/crio-e4d23eb139b7a784ac3e36cdb3dcaac3b90d32e5440fd6c0980dc22fa4be7122 WatchSource:0}: Error finding container e4d23eb139b7a784ac3e36cdb3dcaac3b90d32e5440fd6c0980dc22fa4be7122: Status 404 returned error can't find the container with id e4d23eb139b7a784ac3e36cdb3dcaac3b90d32e5440fd6c0980dc22fa4be7122 Dec 06 14:12:35 crc kubenswrapper[4706]: I1206 14:12:35.948812 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-fzrvx" event={"ID":"7af82ce5-6c72-4654-b677-f517f8885495","Type":"ContainerStarted","Data":"e4d23eb139b7a784ac3e36cdb3dcaac3b90d32e5440fd6c0980dc22fa4be7122"} Dec 06 14:12:36 crc kubenswrapper[4706]: I1206 14:12:36.958405 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-klwkp" event={"ID":"c36a0ce1-c1a6-4529-9579-a48508a688fc","Type":"ContainerStarted","Data":"27674d1e2788e70273f60a460cc1e0d92991def0d648c3ca66dbcab803e71fdd"} Dec 06 14:12:36 crc kubenswrapper[4706]: I1206 14:12:36.975854 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-index-klwkp" podStartSLOduration=1.809489803 podStartE2EDuration="4.975832136s" podCreationTimestamp="2025-12-06 14:12:32 +0000 UTC" firstStartedPulling="2025-12-06 14:12:33.091930867 +0000 UTC m=+995.927723724" lastFinishedPulling="2025-12-06 14:12:36.2582732 +0000 UTC m=+999.094066057" observedRunningTime="2025-12-06 14:12:36.971335057 +0000 UTC m=+999.807127904" watchObservedRunningTime="2025-12-06 14:12:36.975832136 +0000 UTC m=+999.811625003" Dec 06 14:12:38 crc kubenswrapper[4706]: I1206 14:12:38.864540 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2zxns"] Dec 06 14:12:38 crc kubenswrapper[4706]: I1206 14:12:38.866959 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2zxns" Dec 06 14:12:38 crc kubenswrapper[4706]: I1206 14:12:38.882876 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2zxns"] Dec 06 14:12:38 crc kubenswrapper[4706]: I1206 14:12:38.998852 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8d5l\" (UniqueName: \"kubernetes.io/projected/af4a840d-49f7-4c68-a3e7-b5a315781a1b-kube-api-access-l8d5l\") pod \"community-operators-2zxns\" (UID: \"af4a840d-49f7-4c68-a3e7-b5a315781a1b\") " pod="openshift-marketplace/community-operators-2zxns" Dec 06 14:12:38 crc kubenswrapper[4706]: I1206 14:12:38.999185 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af4a840d-49f7-4c68-a3e7-b5a315781a1b-utilities\") pod \"community-operators-2zxns\" (UID: \"af4a840d-49f7-4c68-a3e7-b5a315781a1b\") " pod="openshift-marketplace/community-operators-2zxns" Dec 06 14:12:38 crc kubenswrapper[4706]: I1206 14:12:38.999304 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af4a840d-49f7-4c68-a3e7-b5a315781a1b-catalog-content\") pod \"community-operators-2zxns\" (UID: \"af4a840d-49f7-4c68-a3e7-b5a315781a1b\") " pod="openshift-marketplace/community-operators-2zxns" Dec 06 14:12:39 crc kubenswrapper[4706]: I1206 14:12:39.102434 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8d5l\" (UniqueName: \"kubernetes.io/projected/af4a840d-49f7-4c68-a3e7-b5a315781a1b-kube-api-access-l8d5l\") pod \"community-operators-2zxns\" (UID: \"af4a840d-49f7-4c68-a3e7-b5a315781a1b\") " pod="openshift-marketplace/community-operators-2zxns" Dec 06 14:12:39 crc kubenswrapper[4706]: I1206 14:12:39.102759 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af4a840d-49f7-4c68-a3e7-b5a315781a1b-utilities\") pod \"community-operators-2zxns\" (UID: \"af4a840d-49f7-4c68-a3e7-b5a315781a1b\") " pod="openshift-marketplace/community-operators-2zxns" Dec 06 14:12:39 crc kubenswrapper[4706]: I1206 14:12:39.102853 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af4a840d-49f7-4c68-a3e7-b5a315781a1b-catalog-content\") pod \"community-operators-2zxns\" (UID: \"af4a840d-49f7-4c68-a3e7-b5a315781a1b\") " pod="openshift-marketplace/community-operators-2zxns" Dec 06 14:12:39 crc kubenswrapper[4706]: I1206 14:12:39.103714 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af4a840d-49f7-4c68-a3e7-b5a315781a1b-utilities\") pod \"community-operators-2zxns\" (UID: \"af4a840d-49f7-4c68-a3e7-b5a315781a1b\") " pod="openshift-marketplace/community-operators-2zxns" Dec 06 14:12:39 crc kubenswrapper[4706]: I1206 14:12:39.103792 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af4a840d-49f7-4c68-a3e7-b5a315781a1b-catalog-content\") pod \"community-operators-2zxns\" (UID: \"af4a840d-49f7-4c68-a3e7-b5a315781a1b\") " pod="openshift-marketplace/community-operators-2zxns" Dec 06 14:12:39 crc kubenswrapper[4706]: I1206 14:12:39.140031 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8d5l\" (UniqueName: \"kubernetes.io/projected/af4a840d-49f7-4c68-a3e7-b5a315781a1b-kube-api-access-l8d5l\") pod \"community-operators-2zxns\" (UID: \"af4a840d-49f7-4c68-a3e7-b5a315781a1b\") " pod="openshift-marketplace/community-operators-2zxns" Dec 06 14:12:39 crc kubenswrapper[4706]: I1206 14:12:39.195834 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2zxns" Dec 06 14:12:42 crc kubenswrapper[4706]: I1206 14:12:42.601582 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/cinder-operator-index-klwkp" Dec 06 14:12:42 crc kubenswrapper[4706]: I1206 14:12:42.602204 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-index-klwkp" Dec 06 14:12:42 crc kubenswrapper[4706]: I1206 14:12:42.641689 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/cinder-operator-index-klwkp" Dec 06 14:12:43 crc kubenswrapper[4706]: I1206 14:12:43.041353 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-index-klwkp" Dec 06 14:12:44 crc kubenswrapper[4706]: I1206 14:12:44.580097 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:12:44 crc kubenswrapper[4706]: I1206 14:12:44.580159 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:12:44 crc kubenswrapper[4706]: I1206 14:12:44.580202 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 14:12:44 crc kubenswrapper[4706]: I1206 14:12:44.580751 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fc4d148c623467502d8b194b1ce686e138f01f9dcba1a1b63282c7d88ff885f4"} pod="openshift-machine-config-operator/machine-config-daemon-67v56" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 14:12:44 crc kubenswrapper[4706]: I1206 14:12:44.580802 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" containerID="cri-o://fc4d148c623467502d8b194b1ce686e138f01f9dcba1a1b63282c7d88ff885f4" gracePeriod=600 Dec 06 14:12:45 crc kubenswrapper[4706]: I1206 14:12:45.526792 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf"] Dec 06 14:12:45 crc kubenswrapper[4706]: I1206 14:12:45.530116 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" Dec 06 14:12:45 crc kubenswrapper[4706]: I1206 14:12:45.533660 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-fc6v2" Dec 06 14:12:45 crc kubenswrapper[4706]: I1206 14:12:45.540254 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf"] Dec 06 14:12:45 crc kubenswrapper[4706]: I1206 14:12:45.640194 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96f74c30-7693-4aa2-a171-5739c79cee2a-util\") pod \"be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf\" (UID: \"96f74c30-7693-4aa2-a171-5739c79cee2a\") " pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" Dec 06 14:12:45 crc kubenswrapper[4706]: I1206 14:12:45.640768 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96f74c30-7693-4aa2-a171-5739c79cee2a-bundle\") pod \"be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf\" (UID: \"96f74c30-7693-4aa2-a171-5739c79cee2a\") " pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" Dec 06 14:12:45 crc kubenswrapper[4706]: I1206 14:12:45.640865 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6jjt\" (UniqueName: \"kubernetes.io/projected/96f74c30-7693-4aa2-a171-5739c79cee2a-kube-api-access-w6jjt\") pod \"be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf\" (UID: \"96f74c30-7693-4aa2-a171-5739c79cee2a\") " pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" Dec 06 14:12:45 crc kubenswrapper[4706]: I1206 14:12:45.742512 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6jjt\" (UniqueName: \"kubernetes.io/projected/96f74c30-7693-4aa2-a171-5739c79cee2a-kube-api-access-w6jjt\") pod \"be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf\" (UID: \"96f74c30-7693-4aa2-a171-5739c79cee2a\") " pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" Dec 06 14:12:45 crc kubenswrapper[4706]: I1206 14:12:45.742754 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96f74c30-7693-4aa2-a171-5739c79cee2a-util\") pod \"be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf\" (UID: \"96f74c30-7693-4aa2-a171-5739c79cee2a\") " pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" Dec 06 14:12:45 crc kubenswrapper[4706]: I1206 14:12:45.743548 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96f74c30-7693-4aa2-a171-5739c79cee2a-util\") pod \"be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf\" (UID: \"96f74c30-7693-4aa2-a171-5739c79cee2a\") " pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" Dec 06 14:12:45 crc kubenswrapper[4706]: I1206 14:12:45.743734 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96f74c30-7693-4aa2-a171-5739c79cee2a-bundle\") pod \"be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf\" (UID: \"96f74c30-7693-4aa2-a171-5739c79cee2a\") " pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" Dec 06 14:12:45 crc kubenswrapper[4706]: I1206 14:12:45.744225 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96f74c30-7693-4aa2-a171-5739c79cee2a-bundle\") pod \"be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf\" (UID: \"96f74c30-7693-4aa2-a171-5739c79cee2a\") " pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" Dec 06 14:12:45 crc kubenswrapper[4706]: I1206 14:12:45.777401 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6jjt\" (UniqueName: \"kubernetes.io/projected/96f74c30-7693-4aa2-a171-5739c79cee2a-kube-api-access-w6jjt\") pod \"be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf\" (UID: \"96f74c30-7693-4aa2-a171-5739c79cee2a\") " pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" Dec 06 14:12:45 crc kubenswrapper[4706]: I1206 14:12:45.859581 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" Dec 06 14:12:47 crc kubenswrapper[4706]: I1206 14:12:47.044853 4706 generic.go:334] "Generic (PLEG): container finished" podID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerID="fc4d148c623467502d8b194b1ce686e138f01f9dcba1a1b63282c7d88ff885f4" exitCode=0 Dec 06 14:12:47 crc kubenswrapper[4706]: I1206 14:12:47.044903 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerDied","Data":"fc4d148c623467502d8b194b1ce686e138f01f9dcba1a1b63282c7d88ff885f4"} Dec 06 14:12:47 crc kubenswrapper[4706]: I1206 14:12:47.044962 4706 scope.go:117] "RemoveContainer" containerID="8f8d8c75a654551bee4406e15ba9924159f4ad79dd336b58107e1b41c26cbce8" Dec 06 14:12:47 crc kubenswrapper[4706]: I1206 14:12:47.283756 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2zxns"] Dec 06 14:12:47 crc kubenswrapper[4706]: I1206 14:12:47.566233 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf"] Dec 06 14:12:47 crc kubenswrapper[4706]: W1206 14:12:47.569317 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96f74c30_7693_4aa2_a171_5739c79cee2a.slice/crio-f6d3494a05f58c13438b55832c3530991fc8b468027372bb70bc9eb04109a0c0 WatchSource:0}: Error finding container f6d3494a05f58c13438b55832c3530991fc8b468027372bb70bc9eb04109a0c0: Status 404 returned error can't find the container with id f6d3494a05f58c13438b55832c3530991fc8b468027372bb70bc9eb04109a0c0 Dec 06 14:12:48 crc kubenswrapper[4706]: I1206 14:12:48.055590 4706 generic.go:334] "Generic (PLEG): container finished" podID="96f74c30-7693-4aa2-a171-5739c79cee2a" containerID="5004a92cee42e60bd649ea2769194287aca089e4702405bd3cc825bcc85d5eec" exitCode=0 Dec 06 14:12:48 crc kubenswrapper[4706]: I1206 14:12:48.055679 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" event={"ID":"96f74c30-7693-4aa2-a171-5739c79cee2a","Type":"ContainerDied","Data":"5004a92cee42e60bd649ea2769194287aca089e4702405bd3cc825bcc85d5eec"} Dec 06 14:12:48 crc kubenswrapper[4706]: I1206 14:12:48.056144 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" event={"ID":"96f74c30-7693-4aa2-a171-5739c79cee2a","Type":"ContainerStarted","Data":"f6d3494a05f58c13438b55832c3530991fc8b468027372bb70bc9eb04109a0c0"} Dec 06 14:12:48 crc kubenswrapper[4706]: I1206 14:12:48.059110 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerStarted","Data":"953455e6657a15cff411732f38c90183866bd015ab24e23297c405658c8efab6"} Dec 06 14:12:48 crc kubenswrapper[4706]: I1206 14:12:48.060501 4706 generic.go:334] "Generic (PLEG): container finished" podID="af4a840d-49f7-4c68-a3e7-b5a315781a1b" containerID="a80c347ea3700f3cfd7513f95dd9d061e5d3252bdfeea1adf849dba284041e4a" exitCode=0 Dec 06 14:12:48 crc kubenswrapper[4706]: I1206 14:12:48.060545 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zxns" event={"ID":"af4a840d-49f7-4c68-a3e7-b5a315781a1b","Type":"ContainerDied","Data":"a80c347ea3700f3cfd7513f95dd9d061e5d3252bdfeea1adf849dba284041e4a"} Dec 06 14:12:48 crc kubenswrapper[4706]: I1206 14:12:48.060588 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zxns" event={"ID":"af4a840d-49f7-4c68-a3e7-b5a315781a1b","Type":"ContainerStarted","Data":"dc0748588ca8538c1e0a8c561b8a4290e6061f1661c00ad5c9ec9cf875f078ee"} Dec 06 14:12:48 crc kubenswrapper[4706]: I1206 14:12:48.061862 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-fzrvx" event={"ID":"7af82ce5-6c72-4654-b677-f517f8885495","Type":"ContainerStarted","Data":"f05fd2d6df14194e82d8bd3794000b922adb1c66d3841ffc646e63f50fe9eccb"} Dec 06 14:12:48 crc kubenswrapper[4706]: I1206 14:12:48.151451 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-db-sync-fzrvx" podStartSLOduration=1.957527097 podStartE2EDuration="14.151434408s" podCreationTimestamp="2025-12-06 14:12:34 +0000 UTC" firstStartedPulling="2025-12-06 14:12:35.064339772 +0000 UTC m=+997.900132649" lastFinishedPulling="2025-12-06 14:12:47.258247103 +0000 UTC m=+1010.094039960" observedRunningTime="2025-12-06 14:12:48.14887691 +0000 UTC m=+1010.984669767" watchObservedRunningTime="2025-12-06 14:12:48.151434408 +0000 UTC m=+1010.987227275" Dec 06 14:12:51 crc kubenswrapper[4706]: I1206 14:12:51.102907 4706 generic.go:334] "Generic (PLEG): container finished" podID="96f74c30-7693-4aa2-a171-5739c79cee2a" containerID="e856fc6f0fb78df070ed34276cb3d2465ed9db0bea3b53083e26354b90448d7e" exitCode=0 Dec 06 14:12:51 crc kubenswrapper[4706]: I1206 14:12:51.103011 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" event={"ID":"96f74c30-7693-4aa2-a171-5739c79cee2a","Type":"ContainerDied","Data":"e856fc6f0fb78df070ed34276cb3d2465ed9db0bea3b53083e26354b90448d7e"} Dec 06 14:12:51 crc kubenswrapper[4706]: I1206 14:12:51.105380 4706 generic.go:334] "Generic (PLEG): container finished" podID="af4a840d-49f7-4c68-a3e7-b5a315781a1b" containerID="0fd5b9691b69e6016a0f9cdb87925043365d8ab40778f0eb37af8fec8a0890ca" exitCode=0 Dec 06 14:12:51 crc kubenswrapper[4706]: I1206 14:12:51.105423 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zxns" event={"ID":"af4a840d-49f7-4c68-a3e7-b5a315781a1b","Type":"ContainerDied","Data":"0fd5b9691b69e6016a0f9cdb87925043365d8ab40778f0eb37af8fec8a0890ca"} Dec 06 14:12:52 crc kubenswrapper[4706]: I1206 14:12:52.124365 4706 generic.go:334] "Generic (PLEG): container finished" podID="96f74c30-7693-4aa2-a171-5739c79cee2a" containerID="5ae103d7a9afc14c67fbafd0cc97a7632ce9dfb1edc608eeb7e480ff5f9e3e06" exitCode=0 Dec 06 14:12:52 crc kubenswrapper[4706]: I1206 14:12:52.124454 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" event={"ID":"96f74c30-7693-4aa2-a171-5739c79cee2a","Type":"ContainerDied","Data":"5ae103d7a9afc14c67fbafd0cc97a7632ce9dfb1edc608eeb7e480ff5f9e3e06"} Dec 06 14:12:52 crc kubenswrapper[4706]: I1206 14:12:52.128485 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zxns" event={"ID":"af4a840d-49f7-4c68-a3e7-b5a315781a1b","Type":"ContainerStarted","Data":"5604dc36d945f498d12beff11bcf8dd5415541666ada3eecb057c903c52b5521"} Dec 06 14:12:52 crc kubenswrapper[4706]: I1206 14:12:52.171708 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2zxns" podStartSLOduration=10.487183467 podStartE2EDuration="14.171687847s" podCreationTimestamp="2025-12-06 14:12:38 +0000 UTC" firstStartedPulling="2025-12-06 14:12:48.061573269 +0000 UTC m=+1010.897366126" lastFinishedPulling="2025-12-06 14:12:51.746077639 +0000 UTC m=+1014.581870506" observedRunningTime="2025-12-06 14:12:52.169261693 +0000 UTC m=+1015.005054570" watchObservedRunningTime="2025-12-06 14:12:52.171687847 +0000 UTC m=+1015.007480724" Dec 06 14:12:53 crc kubenswrapper[4706]: I1206 14:12:53.462036 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" Dec 06 14:12:53 crc kubenswrapper[4706]: I1206 14:12:53.556254 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6jjt\" (UniqueName: \"kubernetes.io/projected/96f74c30-7693-4aa2-a171-5739c79cee2a-kube-api-access-w6jjt\") pod \"96f74c30-7693-4aa2-a171-5739c79cee2a\" (UID: \"96f74c30-7693-4aa2-a171-5739c79cee2a\") " Dec 06 14:12:53 crc kubenswrapper[4706]: I1206 14:12:53.556416 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96f74c30-7693-4aa2-a171-5739c79cee2a-util\") pod \"96f74c30-7693-4aa2-a171-5739c79cee2a\" (UID: \"96f74c30-7693-4aa2-a171-5739c79cee2a\") " Dec 06 14:12:53 crc kubenswrapper[4706]: I1206 14:12:53.556468 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96f74c30-7693-4aa2-a171-5739c79cee2a-bundle\") pod \"96f74c30-7693-4aa2-a171-5739c79cee2a\" (UID: \"96f74c30-7693-4aa2-a171-5739c79cee2a\") " Dec 06 14:12:53 crc kubenswrapper[4706]: I1206 14:12:53.557398 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96f74c30-7693-4aa2-a171-5739c79cee2a-bundle" (OuterVolumeSpecName: "bundle") pod "96f74c30-7693-4aa2-a171-5739c79cee2a" (UID: "96f74c30-7693-4aa2-a171-5739c79cee2a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:12:53 crc kubenswrapper[4706]: I1206 14:12:53.561246 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96f74c30-7693-4aa2-a171-5739c79cee2a-kube-api-access-w6jjt" (OuterVolumeSpecName: "kube-api-access-w6jjt") pod "96f74c30-7693-4aa2-a171-5739c79cee2a" (UID: "96f74c30-7693-4aa2-a171-5739c79cee2a"). InnerVolumeSpecName "kube-api-access-w6jjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:12:53 crc kubenswrapper[4706]: I1206 14:12:53.565985 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96f74c30-7693-4aa2-a171-5739c79cee2a-util" (OuterVolumeSpecName: "util") pod "96f74c30-7693-4aa2-a171-5739c79cee2a" (UID: "96f74c30-7693-4aa2-a171-5739c79cee2a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:12:53 crc kubenswrapper[4706]: I1206 14:12:53.658397 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6jjt\" (UniqueName: \"kubernetes.io/projected/96f74c30-7693-4aa2-a171-5739c79cee2a-kube-api-access-w6jjt\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:53 crc kubenswrapper[4706]: I1206 14:12:53.658434 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96f74c30-7693-4aa2-a171-5739c79cee2a-util\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:53 crc kubenswrapper[4706]: I1206 14:12:53.658450 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96f74c30-7693-4aa2-a171-5739c79cee2a-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:54 crc kubenswrapper[4706]: I1206 14:12:54.144863 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" event={"ID":"96f74c30-7693-4aa2-a171-5739c79cee2a","Type":"ContainerDied","Data":"f6d3494a05f58c13438b55832c3530991fc8b468027372bb70bc9eb04109a0c0"} Dec 06 14:12:54 crc kubenswrapper[4706]: I1206 14:12:54.144911 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf" Dec 06 14:12:54 crc kubenswrapper[4706]: I1206 14:12:54.144926 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6d3494a05f58c13438b55832c3530991fc8b468027372bb70bc9eb04109a0c0" Dec 06 14:12:55 crc kubenswrapper[4706]: I1206 14:12:55.156862 4706 generic.go:334] "Generic (PLEG): container finished" podID="7af82ce5-6c72-4654-b677-f517f8885495" containerID="f05fd2d6df14194e82d8bd3794000b922adb1c66d3841ffc646e63f50fe9eccb" exitCode=0 Dec 06 14:12:55 crc kubenswrapper[4706]: I1206 14:12:55.156918 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-fzrvx" event={"ID":"7af82ce5-6c72-4654-b677-f517f8885495","Type":"ContainerDied","Data":"f05fd2d6df14194e82d8bd3794000b922adb1c66d3841ffc646e63f50fe9eccb"} Dec 06 14:12:56 crc kubenswrapper[4706]: I1206 14:12:56.450138 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-fzrvx" Dec 06 14:12:56 crc kubenswrapper[4706]: I1206 14:12:56.515029 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vvbg\" (UniqueName: \"kubernetes.io/projected/7af82ce5-6c72-4654-b677-f517f8885495-kube-api-access-8vvbg\") pod \"7af82ce5-6c72-4654-b677-f517f8885495\" (UID: \"7af82ce5-6c72-4654-b677-f517f8885495\") " Dec 06 14:12:56 crc kubenswrapper[4706]: I1206 14:12:56.515117 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7af82ce5-6c72-4654-b677-f517f8885495-config-data\") pod \"7af82ce5-6c72-4654-b677-f517f8885495\" (UID: \"7af82ce5-6c72-4654-b677-f517f8885495\") " Dec 06 14:12:56 crc kubenswrapper[4706]: I1206 14:12:56.519872 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7af82ce5-6c72-4654-b677-f517f8885495-kube-api-access-8vvbg" (OuterVolumeSpecName: "kube-api-access-8vvbg") pod "7af82ce5-6c72-4654-b677-f517f8885495" (UID: "7af82ce5-6c72-4654-b677-f517f8885495"). InnerVolumeSpecName "kube-api-access-8vvbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:12:56 crc kubenswrapper[4706]: I1206 14:12:56.553166 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7af82ce5-6c72-4654-b677-f517f8885495-config-data" (OuterVolumeSpecName: "config-data") pod "7af82ce5-6c72-4654-b677-f517f8885495" (UID: "7af82ce5-6c72-4654-b677-f517f8885495"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:12:56 crc kubenswrapper[4706]: I1206 14:12:56.616852 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vvbg\" (UniqueName: \"kubernetes.io/projected/7af82ce5-6c72-4654-b677-f517f8885495-kube-api-access-8vvbg\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:56 crc kubenswrapper[4706]: I1206 14:12:56.616886 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7af82ce5-6c72-4654-b677-f517f8885495-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.177628 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-fzrvx" event={"ID":"7af82ce5-6c72-4654-b677-f517f8885495","Type":"ContainerDied","Data":"e4d23eb139b7a784ac3e36cdb3dcaac3b90d32e5440fd6c0980dc22fa4be7122"} Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.178049 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4d23eb139b7a784ac3e36cdb3dcaac3b90d32e5440fd6c0980dc22fa4be7122" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.178150 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-fzrvx" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.413325 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-ls6wk"] Dec 06 14:12:57 crc kubenswrapper[4706]: E1206 14:12:57.414060 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96f74c30-7693-4aa2-a171-5739c79cee2a" containerName="extract" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.414084 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="96f74c30-7693-4aa2-a171-5739c79cee2a" containerName="extract" Dec 06 14:12:57 crc kubenswrapper[4706]: E1206 14:12:57.414100 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7af82ce5-6c72-4654-b677-f517f8885495" containerName="keystone-db-sync" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.414108 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="7af82ce5-6c72-4654-b677-f517f8885495" containerName="keystone-db-sync" Dec 06 14:12:57 crc kubenswrapper[4706]: E1206 14:12:57.414137 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96f74c30-7693-4aa2-a171-5739c79cee2a" containerName="pull" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.414145 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="96f74c30-7693-4aa2-a171-5739c79cee2a" containerName="pull" Dec 06 14:12:57 crc kubenswrapper[4706]: E1206 14:12:57.414157 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96f74c30-7693-4aa2-a171-5739c79cee2a" containerName="util" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.414164 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="96f74c30-7693-4aa2-a171-5739c79cee2a" containerName="util" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.414294 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="7af82ce5-6c72-4654-b677-f517f8885495" containerName="keystone-db-sync" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.414337 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="96f74c30-7693-4aa2-a171-5739c79cee2a" containerName="extract" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.414866 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.417625 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"osp-secret" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.417874 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-scripts" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.418389 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-keystone-dockercfg-5brnv" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.424528 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-ls6wk"] Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.425030 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-config-data" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.425131 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.532725 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-config-data\") pod \"keystone-bootstrap-ls6wk\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.532801 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkzmd\" (UniqueName: \"kubernetes.io/projected/088ef3b1-37c7-4bcd-b508-847327dc1d38-kube-api-access-nkzmd\") pod \"keystone-bootstrap-ls6wk\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.532839 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-fernet-keys\") pod \"keystone-bootstrap-ls6wk\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.532973 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-credential-keys\") pod \"keystone-bootstrap-ls6wk\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.533151 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-scripts\") pod \"keystone-bootstrap-ls6wk\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.636127 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-config-data\") pod \"keystone-bootstrap-ls6wk\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.639113 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-config-data" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.639478 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkzmd\" (UniqueName: \"kubernetes.io/projected/088ef3b1-37c7-4bcd-b508-847327dc1d38-kube-api-access-nkzmd\") pod \"keystone-bootstrap-ls6wk\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.639579 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-fernet-keys\") pod \"keystone-bootstrap-ls6wk\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.639641 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-credential-keys\") pod \"keystone-bootstrap-ls6wk\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.639788 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-scripts\") pod \"keystone-bootstrap-ls6wk\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.644240 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.646372 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-scripts" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.655948 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-scripts\") pod \"keystone-bootstrap-ls6wk\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.656629 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-fernet-keys\") pod \"keystone-bootstrap-ls6wk\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.657978 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-config-data\") pod \"keystone-bootstrap-ls6wk\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.662770 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-credential-keys\") pod \"keystone-bootstrap-ls6wk\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.670396 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkzmd\" (UniqueName: \"kubernetes.io/projected/088ef3b1-37c7-4bcd-b508-847327dc1d38-kube-api-access-nkzmd\") pod \"keystone-bootstrap-ls6wk\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.737987 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-keystone-dockercfg-5brnv" Dec 06 14:12:57 crc kubenswrapper[4706]: I1206 14:12:57.745487 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:12:58 crc kubenswrapper[4706]: I1206 14:12:58.332457 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-ls6wk"] Dec 06 14:12:58 crc kubenswrapper[4706]: I1206 14:12:58.379887 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"osp-secret" Dec 06 14:12:59 crc kubenswrapper[4706]: I1206 14:12:59.195779 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" event={"ID":"088ef3b1-37c7-4bcd-b508-847327dc1d38","Type":"ContainerStarted","Data":"e3008d0f7a13b4ae6babdabf54dcbca49d6ed97057404b8e49886db05c8bec93"} Dec 06 14:12:59 crc kubenswrapper[4706]: I1206 14:12:59.197423 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2zxns" Dec 06 14:12:59 crc kubenswrapper[4706]: I1206 14:12:59.197461 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2zxns" Dec 06 14:12:59 crc kubenswrapper[4706]: I1206 14:12:59.242406 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2zxns" Dec 06 14:13:00 crc kubenswrapper[4706]: I1206 14:13:00.208555 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" event={"ID":"088ef3b1-37c7-4bcd-b508-847327dc1d38","Type":"ContainerStarted","Data":"4dbdd10af7c3a934f9c28fb5839b4d5dde743af583343c6d9a9bf4ba03475274"} Dec 06 14:13:00 crc kubenswrapper[4706]: I1206 14:13:00.265906 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2zxns" Dec 06 14:13:01 crc kubenswrapper[4706]: I1206 14:13:01.236105 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" podStartSLOduration=4.236089179 podStartE2EDuration="4.236089179s" podCreationTimestamp="2025-12-06 14:12:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:13:01.228998592 +0000 UTC m=+1024.064791449" watchObservedRunningTime="2025-12-06 14:13:01.236089179 +0000 UTC m=+1024.071882036" Dec 06 14:13:02 crc kubenswrapper[4706]: I1206 14:13:02.653482 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2zxns"] Dec 06 14:13:02 crc kubenswrapper[4706]: I1206 14:13:02.654120 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2zxns" podUID="af4a840d-49f7-4c68-a3e7-b5a315781a1b" containerName="registry-server" containerID="cri-o://5604dc36d945f498d12beff11bcf8dd5415541666ada3eecb057c903c52b5521" gracePeriod=2 Dec 06 14:13:03 crc kubenswrapper[4706]: I1206 14:13:03.237474 4706 generic.go:334] "Generic (PLEG): container finished" podID="af4a840d-49f7-4c68-a3e7-b5a315781a1b" containerID="5604dc36d945f498d12beff11bcf8dd5415541666ada3eecb057c903c52b5521" exitCode=0 Dec 06 14:13:03 crc kubenswrapper[4706]: I1206 14:13:03.237522 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zxns" event={"ID":"af4a840d-49f7-4c68-a3e7-b5a315781a1b","Type":"ContainerDied","Data":"5604dc36d945f498d12beff11bcf8dd5415541666ada3eecb057c903c52b5521"} Dec 06 14:13:03 crc kubenswrapper[4706]: I1206 14:13:03.452103 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2zxns" Dec 06 14:13:03 crc kubenswrapper[4706]: I1206 14:13:03.522933 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8d5l\" (UniqueName: \"kubernetes.io/projected/af4a840d-49f7-4c68-a3e7-b5a315781a1b-kube-api-access-l8d5l\") pod \"af4a840d-49f7-4c68-a3e7-b5a315781a1b\" (UID: \"af4a840d-49f7-4c68-a3e7-b5a315781a1b\") " Dec 06 14:13:03 crc kubenswrapper[4706]: I1206 14:13:03.523056 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af4a840d-49f7-4c68-a3e7-b5a315781a1b-catalog-content\") pod \"af4a840d-49f7-4c68-a3e7-b5a315781a1b\" (UID: \"af4a840d-49f7-4c68-a3e7-b5a315781a1b\") " Dec 06 14:13:03 crc kubenswrapper[4706]: I1206 14:13:03.523142 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af4a840d-49f7-4c68-a3e7-b5a315781a1b-utilities\") pod \"af4a840d-49f7-4c68-a3e7-b5a315781a1b\" (UID: \"af4a840d-49f7-4c68-a3e7-b5a315781a1b\") " Dec 06 14:13:03 crc kubenswrapper[4706]: I1206 14:13:03.524171 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af4a840d-49f7-4c68-a3e7-b5a315781a1b-utilities" (OuterVolumeSpecName: "utilities") pod "af4a840d-49f7-4c68-a3e7-b5a315781a1b" (UID: "af4a840d-49f7-4c68-a3e7-b5a315781a1b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:13:03 crc kubenswrapper[4706]: I1206 14:13:03.529715 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af4a840d-49f7-4c68-a3e7-b5a315781a1b-kube-api-access-l8d5l" (OuterVolumeSpecName: "kube-api-access-l8d5l") pod "af4a840d-49f7-4c68-a3e7-b5a315781a1b" (UID: "af4a840d-49f7-4c68-a3e7-b5a315781a1b"). InnerVolumeSpecName "kube-api-access-l8d5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:13:03 crc kubenswrapper[4706]: I1206 14:13:03.590001 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af4a840d-49f7-4c68-a3e7-b5a315781a1b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af4a840d-49f7-4c68-a3e7-b5a315781a1b" (UID: "af4a840d-49f7-4c68-a3e7-b5a315781a1b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:13:03 crc kubenswrapper[4706]: I1206 14:13:03.624888 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af4a840d-49f7-4c68-a3e7-b5a315781a1b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 14:13:03 crc kubenswrapper[4706]: I1206 14:13:03.624921 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af4a840d-49f7-4c68-a3e7-b5a315781a1b-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 14:13:03 crc kubenswrapper[4706]: I1206 14:13:03.624933 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8d5l\" (UniqueName: \"kubernetes.io/projected/af4a840d-49f7-4c68-a3e7-b5a315781a1b-kube-api-access-l8d5l\") on node \"crc\" DevicePath \"\"" Dec 06 14:13:04 crc kubenswrapper[4706]: I1206 14:13:04.252548 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zxns" event={"ID":"af4a840d-49f7-4c68-a3e7-b5a315781a1b","Type":"ContainerDied","Data":"dc0748588ca8538c1e0a8c561b8a4290e6061f1661c00ad5c9ec9cf875f078ee"} Dec 06 14:13:04 crc kubenswrapper[4706]: I1206 14:13:04.252639 4706 scope.go:117] "RemoveContainer" containerID="5604dc36d945f498d12beff11bcf8dd5415541666ada3eecb057c903c52b5521" Dec 06 14:13:04 crc kubenswrapper[4706]: I1206 14:13:04.252694 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2zxns" Dec 06 14:13:04 crc kubenswrapper[4706]: I1206 14:13:04.269577 4706 scope.go:117] "RemoveContainer" containerID="0fd5b9691b69e6016a0f9cdb87925043365d8ab40778f0eb37af8fec8a0890ca" Dec 06 14:13:04 crc kubenswrapper[4706]: I1206 14:13:04.306615 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2zxns"] Dec 06 14:13:04 crc kubenswrapper[4706]: I1206 14:13:04.306983 4706 scope.go:117] "RemoveContainer" containerID="a80c347ea3700f3cfd7513f95dd9d061e5d3252bdfeea1adf849dba284041e4a" Dec 06 14:13:04 crc kubenswrapper[4706]: I1206 14:13:04.316202 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2zxns"] Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.522707 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af4a840d-49f7-4c68-a3e7-b5a315781a1b" path="/var/lib/kubelet/pods/af4a840d-49f7-4c68-a3e7-b5a315781a1b/volumes" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.670263 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7"] Dec 06 14:13:05 crc kubenswrapper[4706]: E1206 14:13:05.670530 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af4a840d-49f7-4c68-a3e7-b5a315781a1b" containerName="extract-content" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.670545 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="af4a840d-49f7-4c68-a3e7-b5a315781a1b" containerName="extract-content" Dec 06 14:13:05 crc kubenswrapper[4706]: E1206 14:13:05.670561 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af4a840d-49f7-4c68-a3e7-b5a315781a1b" containerName="extract-utilities" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.670568 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="af4a840d-49f7-4c68-a3e7-b5a315781a1b" containerName="extract-utilities" Dec 06 14:13:05 crc kubenswrapper[4706]: E1206 14:13:05.670583 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af4a840d-49f7-4c68-a3e7-b5a315781a1b" containerName="registry-server" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.670589 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="af4a840d-49f7-4c68-a3e7-b5a315781a1b" containerName="registry-server" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.670693 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="af4a840d-49f7-4c68-a3e7-b5a315781a1b" containerName="registry-server" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.671083 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.673419 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-tf8t4" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.673613 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-service-cert" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.687494 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7"] Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.779955 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8c97961d-a743-4c7d-9bde-f319fea2af90-webhook-cert\") pod \"cinder-operator-controller-manager-65b4f9dcf8-ltdm7\" (UID: \"8c97961d-a743-4c7d-9bde-f319fea2af90\") " pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.780271 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8c97961d-a743-4c7d-9bde-f319fea2af90-apiservice-cert\") pod \"cinder-operator-controller-manager-65b4f9dcf8-ltdm7\" (UID: \"8c97961d-a743-4c7d-9bde-f319fea2af90\") " pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.780326 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp5x8\" (UniqueName: \"kubernetes.io/projected/8c97961d-a743-4c7d-9bde-f319fea2af90-kube-api-access-rp5x8\") pod \"cinder-operator-controller-manager-65b4f9dcf8-ltdm7\" (UID: \"8c97961d-a743-4c7d-9bde-f319fea2af90\") " pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.882163 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8c97961d-a743-4c7d-9bde-f319fea2af90-apiservice-cert\") pod \"cinder-operator-controller-manager-65b4f9dcf8-ltdm7\" (UID: \"8c97961d-a743-4c7d-9bde-f319fea2af90\") " pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.882226 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp5x8\" (UniqueName: \"kubernetes.io/projected/8c97961d-a743-4c7d-9bde-f319fea2af90-kube-api-access-rp5x8\") pod \"cinder-operator-controller-manager-65b4f9dcf8-ltdm7\" (UID: \"8c97961d-a743-4c7d-9bde-f319fea2af90\") " pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.882366 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8c97961d-a743-4c7d-9bde-f319fea2af90-webhook-cert\") pod \"cinder-operator-controller-manager-65b4f9dcf8-ltdm7\" (UID: \"8c97961d-a743-4c7d-9bde-f319fea2af90\") " pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.893062 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8c97961d-a743-4c7d-9bde-f319fea2af90-webhook-cert\") pod \"cinder-operator-controller-manager-65b4f9dcf8-ltdm7\" (UID: \"8c97961d-a743-4c7d-9bde-f319fea2af90\") " pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.904924 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8c97961d-a743-4c7d-9bde-f319fea2af90-apiservice-cert\") pod \"cinder-operator-controller-manager-65b4f9dcf8-ltdm7\" (UID: \"8c97961d-a743-4c7d-9bde-f319fea2af90\") " pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.931495 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp5x8\" (UniqueName: \"kubernetes.io/projected/8c97961d-a743-4c7d-9bde-f319fea2af90-kube-api-access-rp5x8\") pod \"cinder-operator-controller-manager-65b4f9dcf8-ltdm7\" (UID: \"8c97961d-a743-4c7d-9bde-f319fea2af90\") " pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" Dec 06 14:13:05 crc kubenswrapper[4706]: I1206 14:13:05.988230 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" Dec 06 14:13:06 crc kubenswrapper[4706]: I1206 14:13:06.436298 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7"] Dec 06 14:13:06 crc kubenswrapper[4706]: W1206 14:13:06.441795 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c97961d_a743_4c7d_9bde_f319fea2af90.slice/crio-d26e4af2db8eae4bfd2383991852132847f374ce781c66bf03650a0e5abe3b94 WatchSource:0}: Error finding container d26e4af2db8eae4bfd2383991852132847f374ce781c66bf03650a0e5abe3b94: Status 404 returned error can't find the container with id d26e4af2db8eae4bfd2383991852132847f374ce781c66bf03650a0e5abe3b94 Dec 06 14:13:06 crc kubenswrapper[4706]: I1206 14:13:06.445387 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 14:13:07 crc kubenswrapper[4706]: I1206 14:13:07.274501 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" event={"ID":"8c97961d-a743-4c7d-9bde-f319fea2af90","Type":"ContainerStarted","Data":"d26e4af2db8eae4bfd2383991852132847f374ce781c66bf03650a0e5abe3b94"} Dec 06 14:13:08 crc kubenswrapper[4706]: I1206 14:13:08.300127 4706 generic.go:334] "Generic (PLEG): container finished" podID="088ef3b1-37c7-4bcd-b508-847327dc1d38" containerID="4dbdd10af7c3a934f9c28fb5839b4d5dde743af583343c6d9a9bf4ba03475274" exitCode=0 Dec 06 14:13:08 crc kubenswrapper[4706]: I1206 14:13:08.300224 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" event={"ID":"088ef3b1-37c7-4bcd-b508-847327dc1d38","Type":"ContainerDied","Data":"4dbdd10af7c3a934f9c28fb5839b4d5dde743af583343c6d9a9bf4ba03475274"} Dec 06 14:13:09 crc kubenswrapper[4706]: I1206 14:13:09.675998 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:13:09 crc kubenswrapper[4706]: I1206 14:13:09.748952 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-credential-keys\") pod \"088ef3b1-37c7-4bcd-b508-847327dc1d38\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " Dec 06 14:13:09 crc kubenswrapper[4706]: I1206 14:13:09.749003 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-scripts\") pod \"088ef3b1-37c7-4bcd-b508-847327dc1d38\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " Dec 06 14:13:09 crc kubenswrapper[4706]: I1206 14:13:09.749030 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkzmd\" (UniqueName: \"kubernetes.io/projected/088ef3b1-37c7-4bcd-b508-847327dc1d38-kube-api-access-nkzmd\") pod \"088ef3b1-37c7-4bcd-b508-847327dc1d38\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " Dec 06 14:13:09 crc kubenswrapper[4706]: I1206 14:13:09.749057 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-config-data\") pod \"088ef3b1-37c7-4bcd-b508-847327dc1d38\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " Dec 06 14:13:09 crc kubenswrapper[4706]: I1206 14:13:09.749092 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-fernet-keys\") pod \"088ef3b1-37c7-4bcd-b508-847327dc1d38\" (UID: \"088ef3b1-37c7-4bcd-b508-847327dc1d38\") " Dec 06 14:13:09 crc kubenswrapper[4706]: I1206 14:13:09.754854 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/088ef3b1-37c7-4bcd-b508-847327dc1d38-kube-api-access-nkzmd" (OuterVolumeSpecName: "kube-api-access-nkzmd") pod "088ef3b1-37c7-4bcd-b508-847327dc1d38" (UID: "088ef3b1-37c7-4bcd-b508-847327dc1d38"). InnerVolumeSpecName "kube-api-access-nkzmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:13:09 crc kubenswrapper[4706]: I1206 14:13:09.755275 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "088ef3b1-37c7-4bcd-b508-847327dc1d38" (UID: "088ef3b1-37c7-4bcd-b508-847327dc1d38"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:13:09 crc kubenswrapper[4706]: I1206 14:13:09.755431 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "088ef3b1-37c7-4bcd-b508-847327dc1d38" (UID: "088ef3b1-37c7-4bcd-b508-847327dc1d38"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:13:09 crc kubenswrapper[4706]: I1206 14:13:09.756532 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-scripts" (OuterVolumeSpecName: "scripts") pod "088ef3b1-37c7-4bcd-b508-847327dc1d38" (UID: "088ef3b1-37c7-4bcd-b508-847327dc1d38"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:13:09 crc kubenswrapper[4706]: I1206 14:13:09.769337 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-config-data" (OuterVolumeSpecName: "config-data") pod "088ef3b1-37c7-4bcd-b508-847327dc1d38" (UID: "088ef3b1-37c7-4bcd-b508-847327dc1d38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:13:09 crc kubenswrapper[4706]: I1206 14:13:09.850425 4706 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 14:13:09 crc kubenswrapper[4706]: I1206 14:13:09.850464 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:13:09 crc kubenswrapper[4706]: I1206 14:13:09.850475 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkzmd\" (UniqueName: \"kubernetes.io/projected/088ef3b1-37c7-4bcd-b508-847327dc1d38-kube-api-access-nkzmd\") on node \"crc\" DevicePath \"\"" Dec 06 14:13:09 crc kubenswrapper[4706]: I1206 14:13:09.850486 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:13:09 crc kubenswrapper[4706]: I1206 14:13:09.850495 4706 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/088ef3b1-37c7-4bcd-b508-847327dc1d38-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.314605 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" event={"ID":"088ef3b1-37c7-4bcd-b508-847327dc1d38","Type":"ContainerDied","Data":"e3008d0f7a13b4ae6babdabf54dcbca49d6ed97057404b8e49886db05c8bec93"} Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.314898 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3008d0f7a13b4ae6babdabf54dcbca49d6ed97057404b8e49886db05c8bec93" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.314677 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-ls6wk" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.397742 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh"] Dec 06 14:13:10 crc kubenswrapper[4706]: E1206 14:13:10.398071 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="088ef3b1-37c7-4bcd-b508-847327dc1d38" containerName="keystone-bootstrap" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.398098 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="088ef3b1-37c7-4bcd-b508-847327dc1d38" containerName="keystone-bootstrap" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.398250 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="088ef3b1-37c7-4bcd-b508-847327dc1d38" containerName="keystone-bootstrap" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.398764 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.400657 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.400715 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-keystone-dockercfg-5brnv" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.401033 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-config-data" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.401271 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-scripts" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.413491 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh"] Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.459534 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-credential-keys\") pod \"keystone-7cddcf8fd6-knkjh\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.459714 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-fernet-keys\") pod \"keystone-7cddcf8fd6-knkjh\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.459878 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-scripts\") pod \"keystone-7cddcf8fd6-knkjh\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.459973 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbg72\" (UniqueName: \"kubernetes.io/projected/4ddad38d-0652-48ab-a70e-7eca376ba89d-kube-api-access-qbg72\") pod \"keystone-7cddcf8fd6-knkjh\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.460046 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-config-data\") pod \"keystone-7cddcf8fd6-knkjh\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.561880 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-scripts\") pod \"keystone-7cddcf8fd6-knkjh\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.561953 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbg72\" (UniqueName: \"kubernetes.io/projected/4ddad38d-0652-48ab-a70e-7eca376ba89d-kube-api-access-qbg72\") pod \"keystone-7cddcf8fd6-knkjh\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.562007 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-config-data\") pod \"keystone-7cddcf8fd6-knkjh\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.562060 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-credential-keys\") pod \"keystone-7cddcf8fd6-knkjh\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.562139 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-fernet-keys\") pod \"keystone-7cddcf8fd6-knkjh\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.566889 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-fernet-keys\") pod \"keystone-7cddcf8fd6-knkjh\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.566907 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-config-data\") pod \"keystone-7cddcf8fd6-knkjh\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.567374 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-scripts\") pod \"keystone-7cddcf8fd6-knkjh\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.567624 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-credential-keys\") pod \"keystone-7cddcf8fd6-knkjh\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.589364 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbg72\" (UniqueName: \"kubernetes.io/projected/4ddad38d-0652-48ab-a70e-7eca376ba89d-kube-api-access-qbg72\") pod \"keystone-7cddcf8fd6-knkjh\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.711944 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:10 crc kubenswrapper[4706]: I1206 14:13:10.944646 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh"] Dec 06 14:13:10 crc kubenswrapper[4706]: W1206 14:13:10.950998 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ddad38d_0652_48ab_a70e_7eca376ba89d.slice/crio-78cd8cb0e29c83e185cb9381c70f13284fb64e8840537dff26c90a1d2c228e05 WatchSource:0}: Error finding container 78cd8cb0e29c83e185cb9381c70f13284fb64e8840537dff26c90a1d2c228e05: Status 404 returned error can't find the container with id 78cd8cb0e29c83e185cb9381c70f13284fb64e8840537dff26c90a1d2c228e05 Dec 06 14:13:11 crc kubenswrapper[4706]: I1206 14:13:11.321980 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" event={"ID":"4ddad38d-0652-48ab-a70e-7eca376ba89d","Type":"ContainerStarted","Data":"22412e6cb2fd24e10cf3f94722611789ed23e1b4208f9601d256eca63c15b73c"} Dec 06 14:13:11 crc kubenswrapper[4706]: I1206 14:13:11.322347 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:11 crc kubenswrapper[4706]: I1206 14:13:11.322361 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" event={"ID":"4ddad38d-0652-48ab-a70e-7eca376ba89d","Type":"ContainerStarted","Data":"78cd8cb0e29c83e185cb9381c70f13284fb64e8840537dff26c90a1d2c228e05"} Dec 06 14:13:14 crc kubenswrapper[4706]: I1206 14:13:14.352415 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" event={"ID":"8c97961d-a743-4c7d-9bde-f319fea2af90","Type":"ContainerStarted","Data":"ee5b6933fe042a50a48e3b44db8f21a33564dfde007d5a5b105018d966ac9cae"} Dec 06 14:13:14 crc kubenswrapper[4706]: I1206 14:13:14.353177 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" Dec 06 14:13:14 crc kubenswrapper[4706]: I1206 14:13:14.376271 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" podStartSLOduration=2.322090574 podStartE2EDuration="9.376246129s" podCreationTimestamp="2025-12-06 14:13:05 +0000 UTC" firstStartedPulling="2025-12-06 14:13:06.445064357 +0000 UTC m=+1029.280857224" lastFinishedPulling="2025-12-06 14:13:13.499219912 +0000 UTC m=+1036.335012779" observedRunningTime="2025-12-06 14:13:14.371598926 +0000 UTC m=+1037.207391813" watchObservedRunningTime="2025-12-06 14:13:14.376246129 +0000 UTC m=+1037.212038996" Dec 06 14:13:14 crc kubenswrapper[4706]: I1206 14:13:14.379785 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" podStartSLOduration=4.379769752 podStartE2EDuration="4.379769752s" podCreationTimestamp="2025-12-06 14:13:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:13:11.336009905 +0000 UTC m=+1034.171802762" watchObservedRunningTime="2025-12-06 14:13:14.379769752 +0000 UTC m=+1037.215562629" Dec 06 14:13:26 crc kubenswrapper[4706]: I1206 14:13:25.999292 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.151681 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt"] Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.153458 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.157840 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-create-z5f55"] Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.158874 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-z5f55" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.160385 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-db-secret" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.181438 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt"] Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.187907 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-z5f55"] Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.237910 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzgz7\" (UniqueName: \"kubernetes.io/projected/c8e89294-f497-464b-b4d7-3556ce8152a1-kube-api-access-zzgz7\") pod \"cinder-06a5-account-create-update-29vqt\" (UID: \"c8e89294-f497-464b-b4d7-3556ce8152a1\") " pod="cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.237969 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ws75\" (UniqueName: \"kubernetes.io/projected/2fe1a084-66ec-4f61-ad67-bda961120ab6-kube-api-access-8ws75\") pod \"cinder-db-create-z5f55\" (UID: \"2fe1a084-66ec-4f61-ad67-bda961120ab6\") " pod="cinder-kuttl-tests/cinder-db-create-z5f55" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.237991 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8e89294-f497-464b-b4d7-3556ce8152a1-operator-scripts\") pod \"cinder-06a5-account-create-update-29vqt\" (UID: \"c8e89294-f497-464b-b4d7-3556ce8152a1\") " pod="cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.238017 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fe1a084-66ec-4f61-ad67-bda961120ab6-operator-scripts\") pod \"cinder-db-create-z5f55\" (UID: \"2fe1a084-66ec-4f61-ad67-bda961120ab6\") " pod="cinder-kuttl-tests/cinder-db-create-z5f55" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.339393 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzgz7\" (UniqueName: \"kubernetes.io/projected/c8e89294-f497-464b-b4d7-3556ce8152a1-kube-api-access-zzgz7\") pod \"cinder-06a5-account-create-update-29vqt\" (UID: \"c8e89294-f497-464b-b4d7-3556ce8152a1\") " pod="cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.339747 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ws75\" (UniqueName: \"kubernetes.io/projected/2fe1a084-66ec-4f61-ad67-bda961120ab6-kube-api-access-8ws75\") pod \"cinder-db-create-z5f55\" (UID: \"2fe1a084-66ec-4f61-ad67-bda961120ab6\") " pod="cinder-kuttl-tests/cinder-db-create-z5f55" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.339989 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8e89294-f497-464b-b4d7-3556ce8152a1-operator-scripts\") pod \"cinder-06a5-account-create-update-29vqt\" (UID: \"c8e89294-f497-464b-b4d7-3556ce8152a1\") " pod="cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.340217 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fe1a084-66ec-4f61-ad67-bda961120ab6-operator-scripts\") pod \"cinder-db-create-z5f55\" (UID: \"2fe1a084-66ec-4f61-ad67-bda961120ab6\") " pod="cinder-kuttl-tests/cinder-db-create-z5f55" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.341587 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8e89294-f497-464b-b4d7-3556ce8152a1-operator-scripts\") pod \"cinder-06a5-account-create-update-29vqt\" (UID: \"c8e89294-f497-464b-b4d7-3556ce8152a1\") " pod="cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.341618 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fe1a084-66ec-4f61-ad67-bda961120ab6-operator-scripts\") pod \"cinder-db-create-z5f55\" (UID: \"2fe1a084-66ec-4f61-ad67-bda961120ab6\") " pod="cinder-kuttl-tests/cinder-db-create-z5f55" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.373924 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzgz7\" (UniqueName: \"kubernetes.io/projected/c8e89294-f497-464b-b4d7-3556ce8152a1-kube-api-access-zzgz7\") pod \"cinder-06a5-account-create-update-29vqt\" (UID: \"c8e89294-f497-464b-b4d7-3556ce8152a1\") " pod="cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.375526 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ws75\" (UniqueName: \"kubernetes.io/projected/2fe1a084-66ec-4f61-ad67-bda961120ab6-kube-api-access-8ws75\") pod \"cinder-db-create-z5f55\" (UID: \"2fe1a084-66ec-4f61-ad67-bda961120ab6\") " pod="cinder-kuttl-tests/cinder-db-create-z5f55" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.487575 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.494504 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-z5f55" Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.931840 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-z5f55"] Dec 06 14:13:28 crc kubenswrapper[4706]: I1206 14:13:28.999632 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt"] Dec 06 14:13:29 crc kubenswrapper[4706]: I1206 14:13:29.469640 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt" event={"ID":"c8e89294-f497-464b-b4d7-3556ce8152a1","Type":"ContainerStarted","Data":"b3d2693f98db9ba176c3d7acbd4cb1039e3723d4e5843e88fe8edf7af40c5015"} Dec 06 14:13:29 crc kubenswrapper[4706]: I1206 14:13:29.469963 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt" event={"ID":"c8e89294-f497-464b-b4d7-3556ce8152a1","Type":"ContainerStarted","Data":"7308c8090eb88cb18a30651eb8a23990c33f561173ef9cef2ee88b284906b1c2"} Dec 06 14:13:29 crc kubenswrapper[4706]: I1206 14:13:29.471115 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-z5f55" event={"ID":"2fe1a084-66ec-4f61-ad67-bda961120ab6","Type":"ContainerStarted","Data":"06e0cc770ece8cab7566a1e52eda59fad9955394bd977916b8b619b901688275"} Dec 06 14:13:29 crc kubenswrapper[4706]: I1206 14:13:29.471159 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-z5f55" event={"ID":"2fe1a084-66ec-4f61-ad67-bda961120ab6","Type":"ContainerStarted","Data":"8218385cfc93559c5b9f5414f901325ea9ae57cc970edb4245a8108eaa7a2d58"} Dec 06 14:13:29 crc kubenswrapper[4706]: I1206 14:13:29.487461 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt" podStartSLOduration=1.487438649 podStartE2EDuration="1.487438649s" podCreationTimestamp="2025-12-06 14:13:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:13:29.484289546 +0000 UTC m=+1052.320082413" watchObservedRunningTime="2025-12-06 14:13:29.487438649 +0000 UTC m=+1052.323231506" Dec 06 14:13:29 crc kubenswrapper[4706]: I1206 14:13:29.504748 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-db-create-z5f55" podStartSLOduration=1.504726887 podStartE2EDuration="1.504726887s" podCreationTimestamp="2025-12-06 14:13:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:13:29.498292616 +0000 UTC m=+1052.334085473" watchObservedRunningTime="2025-12-06 14:13:29.504726887 +0000 UTC m=+1052.340519744" Dec 06 14:13:32 crc kubenswrapper[4706]: I1206 14:13:32.496748 4706 generic.go:334] "Generic (PLEG): container finished" podID="2fe1a084-66ec-4f61-ad67-bda961120ab6" containerID="06e0cc770ece8cab7566a1e52eda59fad9955394bd977916b8b619b901688275" exitCode=0 Dec 06 14:13:32 crc kubenswrapper[4706]: I1206 14:13:32.496829 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-z5f55" event={"ID":"2fe1a084-66ec-4f61-ad67-bda961120ab6","Type":"ContainerDied","Data":"06e0cc770ece8cab7566a1e52eda59fad9955394bd977916b8b619b901688275"} Dec 06 14:13:33 crc kubenswrapper[4706]: I1206 14:13:33.507053 4706 generic.go:334] "Generic (PLEG): container finished" podID="c8e89294-f497-464b-b4d7-3556ce8152a1" containerID="b3d2693f98db9ba176c3d7acbd4cb1039e3723d4e5843e88fe8edf7af40c5015" exitCode=0 Dec 06 14:13:33 crc kubenswrapper[4706]: I1206 14:13:33.507150 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt" event={"ID":"c8e89294-f497-464b-b4d7-3556ce8152a1","Type":"ContainerDied","Data":"b3d2693f98db9ba176c3d7acbd4cb1039e3723d4e5843e88fe8edf7af40c5015"} Dec 06 14:13:33 crc kubenswrapper[4706]: I1206 14:13:33.847021 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-z5f55" Dec 06 14:13:33 crc kubenswrapper[4706]: I1206 14:13:33.978416 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fe1a084-66ec-4f61-ad67-bda961120ab6-operator-scripts\") pod \"2fe1a084-66ec-4f61-ad67-bda961120ab6\" (UID: \"2fe1a084-66ec-4f61-ad67-bda961120ab6\") " Dec 06 14:13:33 crc kubenswrapper[4706]: I1206 14:13:33.978501 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ws75\" (UniqueName: \"kubernetes.io/projected/2fe1a084-66ec-4f61-ad67-bda961120ab6-kube-api-access-8ws75\") pod \"2fe1a084-66ec-4f61-ad67-bda961120ab6\" (UID: \"2fe1a084-66ec-4f61-ad67-bda961120ab6\") " Dec 06 14:13:33 crc kubenswrapper[4706]: I1206 14:13:33.979635 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fe1a084-66ec-4f61-ad67-bda961120ab6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2fe1a084-66ec-4f61-ad67-bda961120ab6" (UID: "2fe1a084-66ec-4f61-ad67-bda961120ab6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:13:33 crc kubenswrapper[4706]: I1206 14:13:33.986424 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fe1a084-66ec-4f61-ad67-bda961120ab6-kube-api-access-8ws75" (OuterVolumeSpecName: "kube-api-access-8ws75") pod "2fe1a084-66ec-4f61-ad67-bda961120ab6" (UID: "2fe1a084-66ec-4f61-ad67-bda961120ab6"). InnerVolumeSpecName "kube-api-access-8ws75". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:13:34 crc kubenswrapper[4706]: I1206 14:13:34.080141 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fe1a084-66ec-4f61-ad67-bda961120ab6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:13:34 crc kubenswrapper[4706]: I1206 14:13:34.080174 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ws75\" (UniqueName: \"kubernetes.io/projected/2fe1a084-66ec-4f61-ad67-bda961120ab6-kube-api-access-8ws75\") on node \"crc\" DevicePath \"\"" Dec 06 14:13:34 crc kubenswrapper[4706]: I1206 14:13:34.519295 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-z5f55" Dec 06 14:13:34 crc kubenswrapper[4706]: I1206 14:13:34.521463 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-z5f55" event={"ID":"2fe1a084-66ec-4f61-ad67-bda961120ab6","Type":"ContainerDied","Data":"8218385cfc93559c5b9f5414f901325ea9ae57cc970edb4245a8108eaa7a2d58"} Dec 06 14:13:34 crc kubenswrapper[4706]: I1206 14:13:34.521524 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8218385cfc93559c5b9f5414f901325ea9ae57cc970edb4245a8108eaa7a2d58" Dec 06 14:13:34 crc kubenswrapper[4706]: I1206 14:13:34.894427 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt" Dec 06 14:13:34 crc kubenswrapper[4706]: I1206 14:13:34.998528 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzgz7\" (UniqueName: \"kubernetes.io/projected/c8e89294-f497-464b-b4d7-3556ce8152a1-kube-api-access-zzgz7\") pod \"c8e89294-f497-464b-b4d7-3556ce8152a1\" (UID: \"c8e89294-f497-464b-b4d7-3556ce8152a1\") " Dec 06 14:13:34 crc kubenswrapper[4706]: I1206 14:13:34.998648 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8e89294-f497-464b-b4d7-3556ce8152a1-operator-scripts\") pod \"c8e89294-f497-464b-b4d7-3556ce8152a1\" (UID: \"c8e89294-f497-464b-b4d7-3556ce8152a1\") " Dec 06 14:13:34 crc kubenswrapper[4706]: I1206 14:13:34.999480 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8e89294-f497-464b-b4d7-3556ce8152a1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c8e89294-f497-464b-b4d7-3556ce8152a1" (UID: "c8e89294-f497-464b-b4d7-3556ce8152a1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:13:35 crc kubenswrapper[4706]: I1206 14:13:35.006593 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8e89294-f497-464b-b4d7-3556ce8152a1-kube-api-access-zzgz7" (OuterVolumeSpecName: "kube-api-access-zzgz7") pod "c8e89294-f497-464b-b4d7-3556ce8152a1" (UID: "c8e89294-f497-464b-b4d7-3556ce8152a1"). InnerVolumeSpecName "kube-api-access-zzgz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:13:35 crc kubenswrapper[4706]: I1206 14:13:35.101643 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8e89294-f497-464b-b4d7-3556ce8152a1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:13:35 crc kubenswrapper[4706]: I1206 14:13:35.101723 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzgz7\" (UniqueName: \"kubernetes.io/projected/c8e89294-f497-464b-b4d7-3556ce8152a1-kube-api-access-zzgz7\") on node \"crc\" DevicePath \"\"" Dec 06 14:13:35 crc kubenswrapper[4706]: I1206 14:13:35.532191 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt" event={"ID":"c8e89294-f497-464b-b4d7-3556ce8152a1","Type":"ContainerDied","Data":"7308c8090eb88cb18a30651eb8a23990c33f561173ef9cef2ee88b284906b1c2"} Dec 06 14:13:35 crc kubenswrapper[4706]: I1206 14:13:35.532250 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7308c8090eb88cb18a30651eb8a23990c33f561173ef9cef2ee88b284906b1c2" Dec 06 14:13:35 crc kubenswrapper[4706]: I1206 14:13:35.532370 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.524262 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-ndc2g"] Dec 06 14:13:38 crc kubenswrapper[4706]: E1206 14:13:38.524807 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8e89294-f497-464b-b4d7-3556ce8152a1" containerName="mariadb-account-create-update" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.524821 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8e89294-f497-464b-b4d7-3556ce8152a1" containerName="mariadb-account-create-update" Dec 06 14:13:38 crc kubenswrapper[4706]: E1206 14:13:38.524841 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fe1a084-66ec-4f61-ad67-bda961120ab6" containerName="mariadb-database-create" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.524849 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fe1a084-66ec-4f61-ad67-bda961120ab6" containerName="mariadb-database-create" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.524994 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8e89294-f497-464b-b4d7-3556ce8152a1" containerName="mariadb-account-create-update" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.525010 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fe1a084-66ec-4f61-ad67-bda961120ab6" containerName="mariadb-database-create" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.525494 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.527876 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.533186 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-xcp72" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.533207 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.553588 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-ndc2g"] Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.665434 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28t6b\" (UniqueName: \"kubernetes.io/projected/f55d2950-f0b5-4a23-89b3-f8df608db21f-kube-api-access-28t6b\") pod \"cinder-db-sync-ndc2g\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.665495 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-scripts\") pod \"cinder-db-sync-ndc2g\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.665585 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f55d2950-f0b5-4a23-89b3-f8df608db21f-etc-machine-id\") pod \"cinder-db-sync-ndc2g\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.665619 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-db-sync-config-data\") pod \"cinder-db-sync-ndc2g\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.665637 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-config-data\") pod \"cinder-db-sync-ndc2g\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.766855 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f55d2950-f0b5-4a23-89b3-f8df608db21f-etc-machine-id\") pod \"cinder-db-sync-ndc2g\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.766925 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-db-sync-config-data\") pod \"cinder-db-sync-ndc2g\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.766955 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-config-data\") pod \"cinder-db-sync-ndc2g\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.766976 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f55d2950-f0b5-4a23-89b3-f8df608db21f-etc-machine-id\") pod \"cinder-db-sync-ndc2g\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.767000 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28t6b\" (UniqueName: \"kubernetes.io/projected/f55d2950-f0b5-4a23-89b3-f8df608db21f-kube-api-access-28t6b\") pod \"cinder-db-sync-ndc2g\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.767036 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-scripts\") pod \"cinder-db-sync-ndc2g\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.771086 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-db-sync-config-data\") pod \"cinder-db-sync-ndc2g\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.771086 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-scripts\") pod \"cinder-db-sync-ndc2g\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.776170 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-config-data\") pod \"cinder-db-sync-ndc2g\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.792986 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28t6b\" (UniqueName: \"kubernetes.io/projected/f55d2950-f0b5-4a23-89b3-f8df608db21f-kube-api-access-28t6b\") pod \"cinder-db-sync-ndc2g\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:13:38 crc kubenswrapper[4706]: I1206 14:13:38.874110 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:13:39 crc kubenswrapper[4706]: I1206 14:13:39.296261 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-ndc2g"] Dec 06 14:13:39 crc kubenswrapper[4706]: I1206 14:13:39.575920 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" event={"ID":"f55d2950-f0b5-4a23-89b3-f8df608db21f","Type":"ContainerStarted","Data":"ebc61d7717f428ac1355f58827cab8f5b1f16fc16acdf6d42d876cc937f8d78d"} Dec 06 14:13:42 crc kubenswrapper[4706]: I1206 14:13:42.067867 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:13:55 crc kubenswrapper[4706]: I1206 14:13:55.746546 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" event={"ID":"f55d2950-f0b5-4a23-89b3-f8df608db21f","Type":"ContainerStarted","Data":"5e8086afc9353a5159ca0256b01989180a096754f120731aa1e5716bc8a0eec6"} Dec 06 14:13:55 crc kubenswrapper[4706]: I1206 14:13:55.777786 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" podStartSLOduration=2.653685043 podStartE2EDuration="17.777761684s" podCreationTimestamp="2025-12-06 14:13:38 +0000 UTC" firstStartedPulling="2025-12-06 14:13:39.321715642 +0000 UTC m=+1062.157508499" lastFinishedPulling="2025-12-06 14:13:54.445792243 +0000 UTC m=+1077.281585140" observedRunningTime="2025-12-06 14:13:55.767571445 +0000 UTC m=+1078.603364342" watchObservedRunningTime="2025-12-06 14:13:55.777761684 +0000 UTC m=+1078.613554581" Dec 06 14:13:59 crc kubenswrapper[4706]: I1206 14:13:59.783761 4706 generic.go:334] "Generic (PLEG): container finished" podID="f55d2950-f0b5-4a23-89b3-f8df608db21f" containerID="5e8086afc9353a5159ca0256b01989180a096754f120731aa1e5716bc8a0eec6" exitCode=0 Dec 06 14:13:59 crc kubenswrapper[4706]: I1206 14:13:59.784547 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" event={"ID":"f55d2950-f0b5-4a23-89b3-f8df608db21f","Type":"ContainerDied","Data":"5e8086afc9353a5159ca0256b01989180a096754f120731aa1e5716bc8a0eec6"} Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.087411 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.224298 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28t6b\" (UniqueName: \"kubernetes.io/projected/f55d2950-f0b5-4a23-89b3-f8df608db21f-kube-api-access-28t6b\") pod \"f55d2950-f0b5-4a23-89b3-f8df608db21f\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.224373 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-config-data\") pod \"f55d2950-f0b5-4a23-89b3-f8df608db21f\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.224424 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-db-sync-config-data\") pod \"f55d2950-f0b5-4a23-89b3-f8df608db21f\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.224769 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f55d2950-f0b5-4a23-89b3-f8df608db21f-etc-machine-id\") pod \"f55d2950-f0b5-4a23-89b3-f8df608db21f\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.224809 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-scripts\") pod \"f55d2950-f0b5-4a23-89b3-f8df608db21f\" (UID: \"f55d2950-f0b5-4a23-89b3-f8df608db21f\") " Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.225183 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f55d2950-f0b5-4a23-89b3-f8df608db21f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f55d2950-f0b5-4a23-89b3-f8df608db21f" (UID: "f55d2950-f0b5-4a23-89b3-f8df608db21f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.244451 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f55d2950-f0b5-4a23-89b3-f8df608db21f-kube-api-access-28t6b" (OuterVolumeSpecName: "kube-api-access-28t6b") pod "f55d2950-f0b5-4a23-89b3-f8df608db21f" (UID: "f55d2950-f0b5-4a23-89b3-f8df608db21f"). InnerVolumeSpecName "kube-api-access-28t6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.244593 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f55d2950-f0b5-4a23-89b3-f8df608db21f" (UID: "f55d2950-f0b5-4a23-89b3-f8df608db21f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.245527 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-scripts" (OuterVolumeSpecName: "scripts") pod "f55d2950-f0b5-4a23-89b3-f8df608db21f" (UID: "f55d2950-f0b5-4a23-89b3-f8df608db21f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.291423 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-config-data" (OuterVolumeSpecName: "config-data") pod "f55d2950-f0b5-4a23-89b3-f8df608db21f" (UID: "f55d2950-f0b5-4a23-89b3-f8df608db21f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.327540 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28t6b\" (UniqueName: \"kubernetes.io/projected/f55d2950-f0b5-4a23-89b3-f8df608db21f-kube-api-access-28t6b\") on node \"crc\" DevicePath \"\"" Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.327580 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.327598 4706 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.327617 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f55d2950-f0b5-4a23-89b3-f8df608db21f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.327633 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f55d2950-f0b5-4a23-89b3-f8df608db21f-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.800487 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" event={"ID":"f55d2950-f0b5-4a23-89b3-f8df608db21f","Type":"ContainerDied","Data":"ebc61d7717f428ac1355f58827cab8f5b1f16fc16acdf6d42d876cc937f8d78d"} Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.801040 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebc61d7717f428ac1355f58827cab8f5b1f16fc16acdf6d42d876cc937f8d78d" Dec 06 14:14:01 crc kubenswrapper[4706]: I1206 14:14:01.800761 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-ndc2g" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.133899 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 06 14:14:02 crc kubenswrapper[4706]: E1206 14:14:02.134151 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f55d2950-f0b5-4a23-89b3-f8df608db21f" containerName="cinder-db-sync" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.134162 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f55d2950-f0b5-4a23-89b3-f8df608db21f" containerName="cinder-db-sync" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.134270 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f55d2950-f0b5-4a23-89b3-f8df608db21f" containerName="cinder-db-sync" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.134922 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.139172 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scheduler-config-data" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.139355 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.139502 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.142184 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-xcp72" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.158858 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.173176 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.174483 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.179142 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-backup-config-data" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.186821 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.222539 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.223576 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.225247 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-volume-volume1-config-data" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.237245 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241059 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241110 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwqnw\" (UniqueName: \"kubernetes.io/projected/2316de26-2427-42bf-9d27-f84217c2c343-kube-api-access-kwqnw\") pod \"cinder-scheduler-0\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241136 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-config-data\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241159 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-dev\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241191 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-run\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241214 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241245 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-scripts\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241270 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-config-data\") pod \"cinder-scheduler-0\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241295 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241351 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-lib-modules\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241377 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-config-data-custom\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241403 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2316de26-2427-42bf-9d27-f84217c2c343-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241424 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241445 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241473 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlzsj\" (UniqueName: \"kubernetes.io/projected/b77d62e6-8988-4135-a645-14fcdd3d9c92-kube-api-access-hlzsj\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241496 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241516 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-scripts\") pod \"cinder-scheduler-0\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241541 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-sys\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.241582 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-nvme\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.342523 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.342596 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.342733 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-lib-modules\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.342780 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-sys\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.342821 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-config-data-custom\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.342837 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-lib-modules\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.342860 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2316de26-2427-42bf-9d27-f84217c2c343-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.342895 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.342914 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.342932 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.342941 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2316de26-2427-42bf-9d27-f84217c2c343-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.342971 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.342966 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.342991 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlzsj\" (UniqueName: \"kubernetes.io/projected/b77d62e6-8988-4135-a645-14fcdd3d9c92-kube-api-access-hlzsj\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343064 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343182 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-scripts\") pod \"cinder-scheduler-0\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343206 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343231 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-sys\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343259 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-sys\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343274 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343325 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp2w5\" (UniqueName: \"kubernetes.io/projected/49814569-1bcb-4ddd-97db-7115d5ec48d1-kube-api-access-vp2w5\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343393 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343442 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-nvme\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343460 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343489 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343523 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343541 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-dev\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343584 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwqnw\" (UniqueName: \"kubernetes.io/projected/2316de26-2427-42bf-9d27-f84217c2c343-kube-api-access-kwqnw\") pod \"cinder-scheduler-0\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343602 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-config-data\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343624 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-run\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343654 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-dev\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343736 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-dev\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343754 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-run\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343793 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-run\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343807 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343890 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-scripts\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343927 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343949 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.343977 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-config-data\") pod \"cinder-scheduler-0\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.344009 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.344065 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.344116 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-nvme\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.344281 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.344499 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.348145 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-config-data-custom\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.348177 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.349997 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-config-data\") pod \"cinder-scheduler-0\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.351746 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-scripts\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.353056 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-scripts\") pod \"cinder-scheduler-0\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.359002 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-config-data\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.359708 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlzsj\" (UniqueName: \"kubernetes.io/projected/b77d62e6-8988-4135-a645-14fcdd3d9c92-kube-api-access-hlzsj\") pod \"cinder-backup-0\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.361264 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwqnw\" (UniqueName: \"kubernetes.io/projected/2316de26-2427-42bf-9d27-f84217c2c343-kube-api-access-kwqnw\") pod \"cinder-scheduler-0\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.445776 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.445842 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp2w5\" (UniqueName: \"kubernetes.io/projected/49814569-1bcb-4ddd-97db-7115d5ec48d1-kube-api-access-vp2w5\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.445893 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.445933 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.445974 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.445928 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446070 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-dev\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446105 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446108 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446012 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-dev\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446197 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-run\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446268 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446285 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-run\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446297 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446374 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446382 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446377 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446421 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446451 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446479 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446488 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-sys\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446549 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-sys\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446613 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.446615 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.450929 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.454008 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.461167 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.465436 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.469005 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-api-config-data" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.473815 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.481639 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.483128 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp2w5\" (UniqueName: \"kubernetes.io/projected/49814569-1bcb-4ddd-97db-7115d5ec48d1-kube-api-access-vp2w5\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.492087 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.498625 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.540724 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.547618 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-scripts\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.547671 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-config-data\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.547710 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f43169ab-3a35-4d23-8758-9e63360c9844-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.547744 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f43169ab-3a35-4d23-8758-9e63360c9844-logs\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.547763 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v25p\" (UniqueName: \"kubernetes.io/projected/f43169ab-3a35-4d23-8758-9e63360c9844-kube-api-access-2v25p\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.547801 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-config-data-custom\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.649996 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-config-data-custom\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.651769 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-scripts\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.651831 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-config-data\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.651911 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f43169ab-3a35-4d23-8758-9e63360c9844-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.651988 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f43169ab-3a35-4d23-8758-9e63360c9844-logs\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.652012 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v25p\" (UniqueName: \"kubernetes.io/projected/f43169ab-3a35-4d23-8758-9e63360c9844-kube-api-access-2v25p\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.655874 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-config-data-custom\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.655919 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f43169ab-3a35-4d23-8758-9e63360c9844-logs\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.656153 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f43169ab-3a35-4d23-8758-9e63360c9844-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.657972 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-scripts\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.658246 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-config-data\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.674659 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v25p\" (UniqueName: \"kubernetes.io/projected/f43169ab-3a35-4d23-8758-9e63360c9844-kube-api-access-2v25p\") pod \"cinder-api-0\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.839897 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.897962 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 06 14:14:02 crc kubenswrapper[4706]: I1206 14:14:02.969832 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 06 14:14:03 crc kubenswrapper[4706]: I1206 14:14:03.027418 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 06 14:14:03 crc kubenswrapper[4706]: I1206 14:14:03.077701 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:14:03 crc kubenswrapper[4706]: W1206 14:14:03.080995 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf43169ab_3a35_4d23_8758_9e63360c9844.slice/crio-71f84d736187d94e242d3c45f580ef46ebe9c369ad86351e20d1959dda440b57 WatchSource:0}: Error finding container 71f84d736187d94e242d3c45f580ef46ebe9c369ad86351e20d1959dda440b57: Status 404 returned error can't find the container with id 71f84d736187d94e242d3c45f580ef46ebe9c369ad86351e20d1959dda440b57 Dec 06 14:14:03 crc kubenswrapper[4706]: I1206 14:14:03.820515 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"2316de26-2427-42bf-9d27-f84217c2c343","Type":"ContainerStarted","Data":"f2e6eb9099f4a4866e6b6553f8dcb509112971aefa01253ff5cc844b0f5607e4"} Dec 06 14:14:03 crc kubenswrapper[4706]: I1206 14:14:03.824269 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"f43169ab-3a35-4d23-8758-9e63360c9844","Type":"ContainerStarted","Data":"71f84d736187d94e242d3c45f580ef46ebe9c369ad86351e20d1959dda440b57"} Dec 06 14:14:03 crc kubenswrapper[4706]: I1206 14:14:03.829687 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"b77d62e6-8988-4135-a645-14fcdd3d9c92","Type":"ContainerStarted","Data":"fcc0ead26c4fab3c5763716207701c59a263e0602d193a6d9957e84935c58b2d"} Dec 06 14:14:03 crc kubenswrapper[4706]: I1206 14:14:03.831649 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerStarted","Data":"45595cc5fe6d491c706a93f685f0990f79af0d1eb411e6c53ebbad86f2458c54"} Dec 06 14:14:04 crc kubenswrapper[4706]: I1206 14:14:04.839917 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"b77d62e6-8988-4135-a645-14fcdd3d9c92","Type":"ContainerStarted","Data":"6734078a417bc2c57cba55d9833e2b2d9915ce3ce5c679df08017c80123fa482"} Dec 06 14:14:04 crc kubenswrapper[4706]: I1206 14:14:04.840445 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"b77d62e6-8988-4135-a645-14fcdd3d9c92","Type":"ContainerStarted","Data":"98c8c2349e48854c0ef7f4fd6f2a3ac4dbc581f868848b433127270a7dca0cab"} Dec 06 14:14:04 crc kubenswrapper[4706]: I1206 14:14:04.845172 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerStarted","Data":"08372a1be8376cd62917b359b27abccc58333c2baaf42fdf089c289379c6c6c6"} Dec 06 14:14:04 crc kubenswrapper[4706]: I1206 14:14:04.845225 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerStarted","Data":"aef1eb34f160c3bb098d4cafba2fa1a6a5651850be630af9db3b711c3ea27fd0"} Dec 06 14:14:04 crc kubenswrapper[4706]: I1206 14:14:04.856176 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"f43169ab-3a35-4d23-8758-9e63360c9844","Type":"ContainerStarted","Data":"ebc70353af85aec8e06892c24ab6c1d7b5df3da8ea1058d9b32b9b0ae3184e27"} Dec 06 14:14:04 crc kubenswrapper[4706]: I1206 14:14:04.856219 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"f43169ab-3a35-4d23-8758-9e63360c9844","Type":"ContainerStarted","Data":"2f18ca8696222578e2b0d0a3d4517e4945077dc4ebfe9588fab77f913b5e5c27"} Dec 06 14:14:04 crc kubenswrapper[4706]: I1206 14:14:04.856828 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:04 crc kubenswrapper[4706]: I1206 14:14:04.867922 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-0" podStartSLOduration=1.6589016920000001 podStartE2EDuration="2.867907328s" podCreationTimestamp="2025-12-06 14:14:02 +0000 UTC" firstStartedPulling="2025-12-06 14:14:02.973884177 +0000 UTC m=+1085.809677034" lastFinishedPulling="2025-12-06 14:14:04.182889803 +0000 UTC m=+1087.018682670" observedRunningTime="2025-12-06 14:14:04.864334273 +0000 UTC m=+1087.700127130" watchObservedRunningTime="2025-12-06 14:14:04.867907328 +0000 UTC m=+1087.703700175" Dec 06 14:14:04 crc kubenswrapper[4706]: I1206 14:14:04.928014 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podStartSLOduration=1.737232626 podStartE2EDuration="2.927995418s" podCreationTimestamp="2025-12-06 14:14:02 +0000 UTC" firstStartedPulling="2025-12-06 14:14:03.040121091 +0000 UTC m=+1085.875913938" lastFinishedPulling="2025-12-06 14:14:04.230883863 +0000 UTC m=+1087.066676730" observedRunningTime="2025-12-06 14:14:04.900763807 +0000 UTC m=+1087.736556664" watchObservedRunningTime="2025-12-06 14:14:04.927995418 +0000 UTC m=+1087.763788275" Dec 06 14:14:04 crc kubenswrapper[4706]: I1206 14:14:04.935181 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-0" podStartSLOduration=2.935160578 podStartE2EDuration="2.935160578s" podCreationTimestamp="2025-12-06 14:14:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:14:04.925669027 +0000 UTC m=+1087.761461884" watchObservedRunningTime="2025-12-06 14:14:04.935160578 +0000 UTC m=+1087.770953435" Dec 06 14:14:05 crc kubenswrapper[4706]: I1206 14:14:05.865551 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"2316de26-2427-42bf-9d27-f84217c2c343","Type":"ContainerStarted","Data":"1b6ce9a5b2d08ad2637a9d0ee2c9cb361dbb07b4118e62737cfd8efb87ccce59"} Dec 06 14:14:05 crc kubenswrapper[4706]: I1206 14:14:05.866302 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"2316de26-2427-42bf-9d27-f84217c2c343","Type":"ContainerStarted","Data":"32f0fc2b98f942acc226f0168315800abf4fa319ea4460af717ccd743092710c"} Dec 06 14:14:05 crc kubenswrapper[4706]: I1206 14:14:05.906757 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-0" podStartSLOduration=2.57297316 podStartE2EDuration="3.906741199s" podCreationTimestamp="2025-12-06 14:14:02 +0000 UTC" firstStartedPulling="2025-12-06 14:14:02.904598963 +0000 UTC m=+1085.740391820" lastFinishedPulling="2025-12-06 14:14:04.238367002 +0000 UTC m=+1087.074159859" observedRunningTime="2025-12-06 14:14:05.903567195 +0000 UTC m=+1088.739360062" watchObservedRunningTime="2025-12-06 14:14:05.906741199 +0000 UTC m=+1088.742534056" Dec 06 14:14:06 crc kubenswrapper[4706]: I1206 14:14:06.874061 4706 generic.go:334] "Generic (PLEG): container finished" podID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerID="08372a1be8376cd62917b359b27abccc58333c2baaf42fdf089c289379c6c6c6" exitCode=1 Dec 06 14:14:06 crc kubenswrapper[4706]: I1206 14:14:06.874162 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerDied","Data":"08372a1be8376cd62917b359b27abccc58333c2baaf42fdf089c289379c6c6c6"} Dec 06 14:14:06 crc kubenswrapper[4706]: I1206 14:14:06.875276 4706 scope.go:117] "RemoveContainer" containerID="08372a1be8376cd62917b359b27abccc58333c2baaf42fdf089c289379c6c6c6" Dec 06 14:14:07 crc kubenswrapper[4706]: I1206 14:14:07.462741 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:07 crc kubenswrapper[4706]: I1206 14:14:07.493526 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:07 crc kubenswrapper[4706]: I1206 14:14:07.541727 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:07 crc kubenswrapper[4706]: I1206 14:14:07.888767 4706 generic.go:334] "Generic (PLEG): container finished" podID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerID="aef1eb34f160c3bb098d4cafba2fa1a6a5651850be630af9db3b711c3ea27fd0" exitCode=1 Dec 06 14:14:07 crc kubenswrapper[4706]: I1206 14:14:07.888868 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerDied","Data":"aef1eb34f160c3bb098d4cafba2fa1a6a5651850be630af9db3b711c3ea27fd0"} Dec 06 14:14:12 crc kubenswrapper[4706]: I1206 14:14:12.543945 4706 prober.go:107] "Probe failed" probeType="Startup" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="cinder-volume" probeResult="failure" output="Get \"http://10.217.0.90:8080/\": dial tcp 10.217.0.90:8080: connect: connection refused" Dec 06 14:14:12 crc kubenswrapper[4706]: I1206 14:14:12.686781 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:14:12 crc kubenswrapper[4706]: I1206 14:14:12.729047 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:14:12 crc kubenswrapper[4706]: I1206 14:14:12.928376 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerStarted","Data":"830df5d0952f4f1d00a6f2bb4bf20b8e7c0584485030c776d93d76e4a6ff87b7"} Dec 06 14:14:12 crc kubenswrapper[4706]: I1206 14:14:12.929178 4706 scope.go:117] "RemoveContainer" containerID="aef1eb34f160c3bb098d4cafba2fa1a6a5651850be630af9db3b711c3ea27fd0" Dec 06 14:14:13 crc kubenswrapper[4706]: I1206 14:14:13.939155 4706 generic.go:334] "Generic (PLEG): container finished" podID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerID="830df5d0952f4f1d00a6f2bb4bf20b8e7c0584485030c776d93d76e4a6ff87b7" exitCode=1 Dec 06 14:14:13 crc kubenswrapper[4706]: I1206 14:14:13.939201 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerDied","Data":"830df5d0952f4f1d00a6f2bb4bf20b8e7c0584485030c776d93d76e4a6ff87b7"} Dec 06 14:14:13 crc kubenswrapper[4706]: I1206 14:14:13.939836 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerStarted","Data":"ce4ea872f35233c3a5e81c68f247ae01cb1f101d4c22c97b5e3592129ecbc510"} Dec 06 14:14:13 crc kubenswrapper[4706]: I1206 14:14:13.939862 4706 scope.go:117] "RemoveContainer" containerID="08372a1be8376cd62917b359b27abccc58333c2baaf42fdf089c289379c6c6c6" Dec 06 14:14:13 crc kubenswrapper[4706]: I1206 14:14:13.940064 4706 scope.go:117] "RemoveContainer" containerID="830df5d0952f4f1d00a6f2bb4bf20b8e7c0584485030c776d93d76e4a6ff87b7" Dec 06 14:14:13 crc kubenswrapper[4706]: E1206 14:14:13.940470 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" Dec 06 14:14:14 crc kubenswrapper[4706]: I1206 14:14:14.710388 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:14:14 crc kubenswrapper[4706]: I1206 14:14:14.948731 4706 generic.go:334] "Generic (PLEG): container finished" podID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerID="ce4ea872f35233c3a5e81c68f247ae01cb1f101d4c22c97b5e3592129ecbc510" exitCode=1 Dec 06 14:14:14 crc kubenswrapper[4706]: I1206 14:14:14.948777 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerDied","Data":"ce4ea872f35233c3a5e81c68f247ae01cb1f101d4c22c97b5e3592129ecbc510"} Dec 06 14:14:14 crc kubenswrapper[4706]: I1206 14:14:14.948849 4706 scope.go:117] "RemoveContainer" containerID="aef1eb34f160c3bb098d4cafba2fa1a6a5651850be630af9db3b711c3ea27fd0" Dec 06 14:14:14 crc kubenswrapper[4706]: I1206 14:14:14.949389 4706 scope.go:117] "RemoveContainer" containerID="ce4ea872f35233c3a5e81c68f247ae01cb1f101d4c22c97b5e3592129ecbc510" Dec 06 14:14:14 crc kubenswrapper[4706]: I1206 14:14:14.949429 4706 scope.go:117] "RemoveContainer" containerID="830df5d0952f4f1d00a6f2bb4bf20b8e7c0584485030c776d93d76e4a6ff87b7" Dec 06 14:14:14 crc kubenswrapper[4706]: E1206 14:14:14.949751 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" Dec 06 14:14:15 crc kubenswrapper[4706]: I1206 14:14:15.791809 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 06 14:14:15 crc kubenswrapper[4706]: I1206 14:14:15.792967 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:15 crc kubenswrapper[4706]: I1206 14:14:15.803477 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 06 14:14:15 crc kubenswrapper[4706]: I1206 14:14:15.951495 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k5wh\" (UniqueName: \"kubernetes.io/projected/d1c3f32b-16af-439e-8bd4-5f0456026ab2-kube-api-access-4k5wh\") pod \"cinder-scheduler-1\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:15 crc kubenswrapper[4706]: I1206 14:14:15.951591 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-scripts\") pod \"cinder-scheduler-1\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:15 crc kubenswrapper[4706]: I1206 14:14:15.951665 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-config-data\") pod \"cinder-scheduler-1\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:15 crc kubenswrapper[4706]: I1206 14:14:15.951723 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d1c3f32b-16af-439e-8bd4-5f0456026ab2-etc-machine-id\") pod \"cinder-scheduler-1\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:15 crc kubenswrapper[4706]: I1206 14:14:15.951755 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-config-data-custom\") pod \"cinder-scheduler-1\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:15 crc kubenswrapper[4706]: I1206 14:14:15.959670 4706 scope.go:117] "RemoveContainer" containerID="ce4ea872f35233c3a5e81c68f247ae01cb1f101d4c22c97b5e3592129ecbc510" Dec 06 14:14:15 crc kubenswrapper[4706]: I1206 14:14:15.959693 4706 scope.go:117] "RemoveContainer" containerID="830df5d0952f4f1d00a6f2bb4bf20b8e7c0584485030c776d93d76e4a6ff87b7" Dec 06 14:14:15 crc kubenswrapper[4706]: E1206 14:14:15.959897 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" Dec 06 14:14:16 crc kubenswrapper[4706]: I1206 14:14:16.052746 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-scripts\") pod \"cinder-scheduler-1\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:16 crc kubenswrapper[4706]: I1206 14:14:16.052830 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-config-data\") pod \"cinder-scheduler-1\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:16 crc kubenswrapper[4706]: I1206 14:14:16.052880 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d1c3f32b-16af-439e-8bd4-5f0456026ab2-etc-machine-id\") pod \"cinder-scheduler-1\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:16 crc kubenswrapper[4706]: I1206 14:14:16.052912 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-config-data-custom\") pod \"cinder-scheduler-1\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:16 crc kubenswrapper[4706]: I1206 14:14:16.052990 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k5wh\" (UniqueName: \"kubernetes.io/projected/d1c3f32b-16af-439e-8bd4-5f0456026ab2-kube-api-access-4k5wh\") pod \"cinder-scheduler-1\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:16 crc kubenswrapper[4706]: I1206 14:14:16.053259 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d1c3f32b-16af-439e-8bd4-5f0456026ab2-etc-machine-id\") pod \"cinder-scheduler-1\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:16 crc kubenswrapper[4706]: I1206 14:14:16.060907 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-config-data-custom\") pod \"cinder-scheduler-1\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:16 crc kubenswrapper[4706]: I1206 14:14:16.060966 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-scripts\") pod \"cinder-scheduler-1\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:16 crc kubenswrapper[4706]: I1206 14:14:16.061938 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-config-data\") pod \"cinder-scheduler-1\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:16 crc kubenswrapper[4706]: I1206 14:14:16.089586 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k5wh\" (UniqueName: \"kubernetes.io/projected/d1c3f32b-16af-439e-8bd4-5f0456026ab2-kube-api-access-4k5wh\") pod \"cinder-scheduler-1\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:16 crc kubenswrapper[4706]: I1206 14:14:16.113401 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:16 crc kubenswrapper[4706]: I1206 14:14:16.577754 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 06 14:14:16 crc kubenswrapper[4706]: I1206 14:14:16.975557 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"d1c3f32b-16af-439e-8bd4-5f0456026ab2","Type":"ContainerStarted","Data":"01d56870694276b246cce49bb74ddb1c37e41fe66944f1a3a3ba7b24110050b0"} Dec 06 14:14:17 crc kubenswrapper[4706]: I1206 14:14:17.541378 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:17 crc kubenswrapper[4706]: I1206 14:14:17.542078 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:17 crc kubenswrapper[4706]: I1206 14:14:17.542204 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:17 crc kubenswrapper[4706]: I1206 14:14:17.543297 4706 scope.go:117] "RemoveContainer" containerID="ce4ea872f35233c3a5e81c68f247ae01cb1f101d4c22c97b5e3592129ecbc510" Dec 06 14:14:17 crc kubenswrapper[4706]: I1206 14:14:17.543366 4706 scope.go:117] "RemoveContainer" containerID="830df5d0952f4f1d00a6f2bb4bf20b8e7c0584485030c776d93d76e4a6ff87b7" Dec 06 14:14:17 crc kubenswrapper[4706]: E1206 14:14:17.543750 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" Dec 06 14:14:17 crc kubenswrapper[4706]: I1206 14:14:17.988957 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"d1c3f32b-16af-439e-8bd4-5f0456026ab2","Type":"ContainerStarted","Data":"637224e34278d3a10c7592c1693ec61a1de040d73b1e8e132faeb41a4def20f4"} Dec 06 14:14:17 crc kubenswrapper[4706]: I1206 14:14:17.989839 4706 scope.go:117] "RemoveContainer" containerID="ce4ea872f35233c3a5e81c68f247ae01cb1f101d4c22c97b5e3592129ecbc510" Dec 06 14:14:17 crc kubenswrapper[4706]: I1206 14:14:17.989868 4706 scope.go:117] "RemoveContainer" containerID="830df5d0952f4f1d00a6f2bb4bf20b8e7c0584485030c776d93d76e4a6ff87b7" Dec 06 14:14:17 crc kubenswrapper[4706]: E1206 14:14:17.990267 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" Dec 06 14:14:19 crc kubenswrapper[4706]: I1206 14:14:19.001940 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"d1c3f32b-16af-439e-8bd4-5f0456026ab2","Type":"ContainerStarted","Data":"fa604fcdb78b97f8eb88e3215d9c8cc3b6a4591e4cb215914ecedac813058336"} Dec 06 14:14:19 crc kubenswrapper[4706]: I1206 14:14:19.057948 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-1" podStartSLOduration=4.057923192 podStartE2EDuration="4.057923192s" podCreationTimestamp="2025-12-06 14:14:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:14:19.044983109 +0000 UTC m=+1101.880775996" watchObservedRunningTime="2025-12-06 14:14:19.057923192 +0000 UTC m=+1101.893716089" Dec 06 14:14:21 crc kubenswrapper[4706]: I1206 14:14:21.114556 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.307519 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.394334 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.396480 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.414905 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.524331 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-scripts\") pod \"cinder-scheduler-2\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.524435 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7zlp\" (UniqueName: \"kubernetes.io/projected/e871919b-dc73-4ac8-a8fc-326200c509f9-kube-api-access-g7zlp\") pod \"cinder-scheduler-2\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.524559 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-config-data-custom\") pod \"cinder-scheduler-2\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.524692 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-config-data\") pod \"cinder-scheduler-2\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.524751 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e871919b-dc73-4ac8-a8fc-326200c509f9-etc-machine-id\") pod \"cinder-scheduler-2\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.626676 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7zlp\" (UniqueName: \"kubernetes.io/projected/e871919b-dc73-4ac8-a8fc-326200c509f9-kube-api-access-g7zlp\") pod \"cinder-scheduler-2\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.626760 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-config-data-custom\") pod \"cinder-scheduler-2\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.626845 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-config-data\") pod \"cinder-scheduler-2\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.626911 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e871919b-dc73-4ac8-a8fc-326200c509f9-etc-machine-id\") pod \"cinder-scheduler-2\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.626963 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-scripts\") pod \"cinder-scheduler-2\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.627085 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e871919b-dc73-4ac8-a8fc-326200c509f9-etc-machine-id\") pod \"cinder-scheduler-2\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.632855 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-config-data\") pod \"cinder-scheduler-2\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.633600 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-config-data-custom\") pod \"cinder-scheduler-2\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.633922 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-scripts\") pod \"cinder-scheduler-2\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.647823 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7zlp\" (UniqueName: \"kubernetes.io/projected/e871919b-dc73-4ac8-a8fc-326200c509f9-kube-api-access-g7zlp\") pod \"cinder-scheduler-2\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:26 crc kubenswrapper[4706]: I1206 14:14:26.742752 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:27 crc kubenswrapper[4706]: I1206 14:14:27.089138 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 06 14:14:28 crc kubenswrapper[4706]: I1206 14:14:28.077689 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"e871919b-dc73-4ac8-a8fc-326200c509f9","Type":"ContainerStarted","Data":"837b67c87a76d1ac51450a0589505f9394697e7744e76b91e04efd4c6f8f714e"} Dec 06 14:14:28 crc kubenswrapper[4706]: I1206 14:14:28.077913 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"e871919b-dc73-4ac8-a8fc-326200c509f9","Type":"ContainerStarted","Data":"e28793e335c3d4b4896126bf927ffc14770f8131e52a9747077ea2fdf3529a93"} Dec 06 14:14:29 crc kubenswrapper[4706]: I1206 14:14:29.085837 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"e871919b-dc73-4ac8-a8fc-326200c509f9","Type":"ContainerStarted","Data":"1b9b1a92aa6bbcd830e778d9246ef77d3833b83d10e54814bc8ce483a596ff0b"} Dec 06 14:14:29 crc kubenswrapper[4706]: I1206 14:14:29.131706 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-2" podStartSLOduration=3.131681096 podStartE2EDuration="3.131681096s" podCreationTimestamp="2025-12-06 14:14:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:14:29.126213861 +0000 UTC m=+1111.962006728" watchObservedRunningTime="2025-12-06 14:14:29.131681096 +0000 UTC m=+1111.967473963" Dec 06 14:14:29 crc kubenswrapper[4706]: I1206 14:14:29.513163 4706 scope.go:117] "RemoveContainer" containerID="ce4ea872f35233c3a5e81c68f247ae01cb1f101d4c22c97b5e3592129ecbc510" Dec 06 14:14:29 crc kubenswrapper[4706]: I1206 14:14:29.513203 4706 scope.go:117] "RemoveContainer" containerID="830df5d0952f4f1d00a6f2bb4bf20b8e7c0584485030c776d93d76e4a6ff87b7" Dec 06 14:14:30 crc kubenswrapper[4706]: I1206 14:14:30.134651 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerStarted","Data":"eacfd1e4a7285c5d5131f1219a33c447e2090f9c230e3d482b8c62a51443c54d"} Dec 06 14:14:31 crc kubenswrapper[4706]: I1206 14:14:31.146820 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerStarted","Data":"00e2f85b4d8d2de959364092eb69e66b15126a02a2a6fedd5f3ad6a761ae3ffa"} Dec 06 14:14:31 crc kubenswrapper[4706]: I1206 14:14:31.744223 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:32 crc kubenswrapper[4706]: I1206 14:14:32.158031 4706 generic.go:334] "Generic (PLEG): container finished" podID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerID="00e2f85b4d8d2de959364092eb69e66b15126a02a2a6fedd5f3ad6a761ae3ffa" exitCode=1 Dec 06 14:14:32 crc kubenswrapper[4706]: I1206 14:14:32.158135 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerDied","Data":"00e2f85b4d8d2de959364092eb69e66b15126a02a2a6fedd5f3ad6a761ae3ffa"} Dec 06 14:14:32 crc kubenswrapper[4706]: I1206 14:14:32.158294 4706 scope.go:117] "RemoveContainer" containerID="830df5d0952f4f1d00a6f2bb4bf20b8e7c0584485030c776d93d76e4a6ff87b7" Dec 06 14:14:32 crc kubenswrapper[4706]: I1206 14:14:32.158523 4706 scope.go:117] "RemoveContainer" containerID="00e2f85b4d8d2de959364092eb69e66b15126a02a2a6fedd5f3ad6a761ae3ffa" Dec 06 14:14:32 crc kubenswrapper[4706]: E1206 14:14:32.158708 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" Dec 06 14:14:32 crc kubenswrapper[4706]: I1206 14:14:32.542014 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:33 crc kubenswrapper[4706]: I1206 14:14:33.168152 4706 generic.go:334] "Generic (PLEG): container finished" podID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerID="eacfd1e4a7285c5d5131f1219a33c447e2090f9c230e3d482b8c62a51443c54d" exitCode=1 Dec 06 14:14:33 crc kubenswrapper[4706]: I1206 14:14:33.168206 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerDied","Data":"eacfd1e4a7285c5d5131f1219a33c447e2090f9c230e3d482b8c62a51443c54d"} Dec 06 14:14:33 crc kubenswrapper[4706]: I1206 14:14:33.168273 4706 scope.go:117] "RemoveContainer" containerID="ce4ea872f35233c3a5e81c68f247ae01cb1f101d4c22c97b5e3592129ecbc510" Dec 06 14:14:33 crc kubenswrapper[4706]: I1206 14:14:33.169259 4706 scope.go:117] "RemoveContainer" containerID="eacfd1e4a7285c5d5131f1219a33c447e2090f9c230e3d482b8c62a51443c54d" Dec 06 14:14:33 crc kubenswrapper[4706]: I1206 14:14:33.169394 4706 scope.go:117] "RemoveContainer" containerID="00e2f85b4d8d2de959364092eb69e66b15126a02a2a6fedd5f3ad6a761ae3ffa" Dec 06 14:14:33 crc kubenswrapper[4706]: E1206 14:14:33.170248 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" Dec 06 14:14:34 crc kubenswrapper[4706]: I1206 14:14:34.184036 4706 scope.go:117] "RemoveContainer" containerID="eacfd1e4a7285c5d5131f1219a33c447e2090f9c230e3d482b8c62a51443c54d" Dec 06 14:14:34 crc kubenswrapper[4706]: I1206 14:14:34.184442 4706 scope.go:117] "RemoveContainer" containerID="00e2f85b4d8d2de959364092eb69e66b15126a02a2a6fedd5f3ad6a761ae3ffa" Dec 06 14:14:34 crc kubenswrapper[4706]: E1206 14:14:34.184845 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" Dec 06 14:14:35 crc kubenswrapper[4706]: I1206 14:14:35.542046 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:35 crc kubenswrapper[4706]: I1206 14:14:35.543046 4706 scope.go:117] "RemoveContainer" containerID="eacfd1e4a7285c5d5131f1219a33c447e2090f9c230e3d482b8c62a51443c54d" Dec 06 14:14:35 crc kubenswrapper[4706]: I1206 14:14:35.543076 4706 scope.go:117] "RemoveContainer" containerID="00e2f85b4d8d2de959364092eb69e66b15126a02a2a6fedd5f3ad6a761ae3ffa" Dec 06 14:14:35 crc kubenswrapper[4706]: E1206 14:14:35.543461 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" Dec 06 14:14:36 crc kubenswrapper[4706]: I1206 14:14:36.951501 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:37 crc kubenswrapper[4706]: I1206 14:14:37.376742 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 06 14:14:37 crc kubenswrapper[4706]: I1206 14:14:37.377304 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-2" podUID="e871919b-dc73-4ac8-a8fc-326200c509f9" containerName="cinder-scheduler" containerID="cri-o://837b67c87a76d1ac51450a0589505f9394697e7744e76b91e04efd4c6f8f714e" gracePeriod=30 Dec 06 14:14:37 crc kubenswrapper[4706]: I1206 14:14:37.377415 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-2" podUID="e871919b-dc73-4ac8-a8fc-326200c509f9" containerName="probe" containerID="cri-o://1b9b1a92aa6bbcd830e778d9246ef77d3833b83d10e54814bc8ce483a596ff0b" gracePeriod=30 Dec 06 14:14:37 crc kubenswrapper[4706]: I1206 14:14:37.541965 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:14:37 crc kubenswrapper[4706]: I1206 14:14:37.543884 4706 scope.go:117] "RemoveContainer" containerID="eacfd1e4a7285c5d5131f1219a33c447e2090f9c230e3d482b8c62a51443c54d" Dec 06 14:14:37 crc kubenswrapper[4706]: I1206 14:14:37.543932 4706 scope.go:117] "RemoveContainer" containerID="00e2f85b4d8d2de959364092eb69e66b15126a02a2a6fedd5f3ad6a761ae3ffa" Dec 06 14:14:37 crc kubenswrapper[4706]: E1206 14:14:37.544837 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" Dec 06 14:14:38 crc kubenswrapper[4706]: I1206 14:14:38.220540 4706 generic.go:334] "Generic (PLEG): container finished" podID="e871919b-dc73-4ac8-a8fc-326200c509f9" containerID="1b9b1a92aa6bbcd830e778d9246ef77d3833b83d10e54814bc8ce483a596ff0b" exitCode=0 Dec 06 14:14:38 crc kubenswrapper[4706]: I1206 14:14:38.220610 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"e871919b-dc73-4ac8-a8fc-326200c509f9","Type":"ContainerDied","Data":"1b9b1a92aa6bbcd830e778d9246ef77d3833b83d10e54814bc8ce483a596ff0b"} Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.231384 4706 generic.go:334] "Generic (PLEG): container finished" podID="e871919b-dc73-4ac8-a8fc-326200c509f9" containerID="837b67c87a76d1ac51450a0589505f9394697e7744e76b91e04efd4c6f8f714e" exitCode=0 Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.231429 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"e871919b-dc73-4ac8-a8fc-326200c509f9","Type":"ContainerDied","Data":"837b67c87a76d1ac51450a0589505f9394697e7744e76b91e04efd4c6f8f714e"} Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.333054 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.452537 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e871919b-dc73-4ac8-a8fc-326200c509f9-etc-machine-id\") pod \"e871919b-dc73-4ac8-a8fc-326200c509f9\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.452822 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-config-data-custom\") pod \"e871919b-dc73-4ac8-a8fc-326200c509f9\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.452988 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-scripts\") pod \"e871919b-dc73-4ac8-a8fc-326200c509f9\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.453162 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-config-data\") pod \"e871919b-dc73-4ac8-a8fc-326200c509f9\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.453329 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7zlp\" (UniqueName: \"kubernetes.io/projected/e871919b-dc73-4ac8-a8fc-326200c509f9-kube-api-access-g7zlp\") pod \"e871919b-dc73-4ac8-a8fc-326200c509f9\" (UID: \"e871919b-dc73-4ac8-a8fc-326200c509f9\") " Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.452656 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e871919b-dc73-4ac8-a8fc-326200c509f9-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e871919b-dc73-4ac8-a8fc-326200c509f9" (UID: "e871919b-dc73-4ac8-a8fc-326200c509f9"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.453940 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e871919b-dc73-4ac8-a8fc-326200c509f9-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.457714 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e871919b-dc73-4ac8-a8fc-326200c509f9" (UID: "e871919b-dc73-4ac8-a8fc-326200c509f9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.458416 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-scripts" (OuterVolumeSpecName: "scripts") pod "e871919b-dc73-4ac8-a8fc-326200c509f9" (UID: "e871919b-dc73-4ac8-a8fc-326200c509f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.460529 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e871919b-dc73-4ac8-a8fc-326200c509f9-kube-api-access-g7zlp" (OuterVolumeSpecName: "kube-api-access-g7zlp") pod "e871919b-dc73-4ac8-a8fc-326200c509f9" (UID: "e871919b-dc73-4ac8-a8fc-326200c509f9"). InnerVolumeSpecName "kube-api-access-g7zlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.522946 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-config-data" (OuterVolumeSpecName: "config-data") pod "e871919b-dc73-4ac8-a8fc-326200c509f9" (UID: "e871919b-dc73-4ac8-a8fc-326200c509f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.555251 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.555375 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.555387 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e871919b-dc73-4ac8-a8fc-326200c509f9-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:14:39 crc kubenswrapper[4706]: I1206 14:14:39.555396 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7zlp\" (UniqueName: \"kubernetes.io/projected/e871919b-dc73-4ac8-a8fc-326200c509f9-kube-api-access-g7zlp\") on node \"crc\" DevicePath \"\"" Dec 06 14:14:40 crc kubenswrapper[4706]: I1206 14:14:40.244531 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"e871919b-dc73-4ac8-a8fc-326200c509f9","Type":"ContainerDied","Data":"e28793e335c3d4b4896126bf927ffc14770f8131e52a9747077ea2fdf3529a93"} Dec 06 14:14:40 crc kubenswrapper[4706]: I1206 14:14:40.245174 4706 scope.go:117] "RemoveContainer" containerID="1b9b1a92aa6bbcd830e778d9246ef77d3833b83d10e54814bc8ce483a596ff0b" Dec 06 14:14:40 crc kubenswrapper[4706]: I1206 14:14:40.244606 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 06 14:14:40 crc kubenswrapper[4706]: I1206 14:14:40.268398 4706 scope.go:117] "RemoveContainer" containerID="837b67c87a76d1ac51450a0589505f9394697e7744e76b91e04efd4c6f8f714e" Dec 06 14:14:40 crc kubenswrapper[4706]: I1206 14:14:40.273412 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 06 14:14:40 crc kubenswrapper[4706]: I1206 14:14:40.280001 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 06 14:14:40 crc kubenswrapper[4706]: I1206 14:14:40.286388 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 06 14:14:40 crc kubenswrapper[4706]: I1206 14:14:40.286902 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-1" podUID="d1c3f32b-16af-439e-8bd4-5f0456026ab2" containerName="cinder-scheduler" containerID="cri-o://637224e34278d3a10c7592c1693ec61a1de040d73b1e8e132faeb41a4def20f4" gracePeriod=30 Dec 06 14:14:40 crc kubenswrapper[4706]: I1206 14:14:40.286985 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-1" podUID="d1c3f32b-16af-439e-8bd4-5f0456026ab2" containerName="probe" containerID="cri-o://fa604fcdb78b97f8eb88e3215d9c8cc3b6a4591e4cb215914ecedac813058336" gracePeriod=30 Dec 06 14:14:41 crc kubenswrapper[4706]: I1206 14:14:41.255469 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1c3f32b-16af-439e-8bd4-5f0456026ab2" containerID="fa604fcdb78b97f8eb88e3215d9c8cc3b6a4591e4cb215914ecedac813058336" exitCode=0 Dec 06 14:14:41 crc kubenswrapper[4706]: I1206 14:14:41.255553 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"d1c3f32b-16af-439e-8bd4-5f0456026ab2","Type":"ContainerDied","Data":"fa604fcdb78b97f8eb88e3215d9c8cc3b6a4591e4cb215914ecedac813058336"} Dec 06 14:14:41 crc kubenswrapper[4706]: I1206 14:14:41.523469 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e871919b-dc73-4ac8-a8fc-326200c509f9" path="/var/lib/kubelet/pods/e871919b-dc73-4ac8-a8fc-326200c509f9/volumes" Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.283884 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1c3f32b-16af-439e-8bd4-5f0456026ab2" containerID="637224e34278d3a10c7592c1693ec61a1de040d73b1e8e132faeb41a4def20f4" exitCode=0 Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.283943 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"d1c3f32b-16af-439e-8bd4-5f0456026ab2","Type":"ContainerDied","Data":"637224e34278d3a10c7592c1693ec61a1de040d73b1e8e132faeb41a4def20f4"} Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.355489 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.534399 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4k5wh\" (UniqueName: \"kubernetes.io/projected/d1c3f32b-16af-439e-8bd4-5f0456026ab2-kube-api-access-4k5wh\") pod \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.534478 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-scripts\") pod \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.534523 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-config-data-custom\") pod \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.534573 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-config-data\") pod \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.534619 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d1c3f32b-16af-439e-8bd4-5f0456026ab2-etc-machine-id\") pod \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\" (UID: \"d1c3f32b-16af-439e-8bd4-5f0456026ab2\") " Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.536418 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1c3f32b-16af-439e-8bd4-5f0456026ab2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d1c3f32b-16af-439e-8bd4-5f0456026ab2" (UID: "d1c3f32b-16af-439e-8bd4-5f0456026ab2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.541508 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d1c3f32b-16af-439e-8bd4-5f0456026ab2" (UID: "d1c3f32b-16af-439e-8bd4-5f0456026ab2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.541678 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1c3f32b-16af-439e-8bd4-5f0456026ab2-kube-api-access-4k5wh" (OuterVolumeSpecName: "kube-api-access-4k5wh") pod "d1c3f32b-16af-439e-8bd4-5f0456026ab2" (UID: "d1c3f32b-16af-439e-8bd4-5f0456026ab2"). InnerVolumeSpecName "kube-api-access-4k5wh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.542429 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-scripts" (OuterVolumeSpecName: "scripts") pod "d1c3f32b-16af-439e-8bd4-5f0456026ab2" (UID: "d1c3f32b-16af-439e-8bd4-5f0456026ab2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.626505 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-config-data" (OuterVolumeSpecName: "config-data") pod "d1c3f32b-16af-439e-8bd4-5f0456026ab2" (UID: "d1c3f32b-16af-439e-8bd4-5f0456026ab2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.635617 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4k5wh\" (UniqueName: \"kubernetes.io/projected/d1c3f32b-16af-439e-8bd4-5f0456026ab2-kube-api-access-4k5wh\") on node \"crc\" DevicePath \"\"" Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.635642 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.635654 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.635664 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1c3f32b-16af-439e-8bd4-5f0456026ab2-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:14:44 crc kubenswrapper[4706]: I1206 14:14:44.635674 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d1c3f32b-16af-439e-8bd4-5f0456026ab2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.295257 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"d1c3f32b-16af-439e-8bd4-5f0456026ab2","Type":"ContainerDied","Data":"01d56870694276b246cce49bb74ddb1c37e41fe66944f1a3a3ba7b24110050b0"} Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.295656 4706 scope.go:117] "RemoveContainer" containerID="fa604fcdb78b97f8eb88e3215d9c8cc3b6a4591e4cb215914ecedac813058336" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.295397 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.329108 4706 scope.go:117] "RemoveContainer" containerID="637224e34278d3a10c7592c1693ec61a1de040d73b1e8e132faeb41a4def20f4" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.353402 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.363360 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.530573 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1c3f32b-16af-439e-8bd4-5f0456026ab2" path="/var/lib/kubelet/pods/d1c3f32b-16af-439e-8bd4-5f0456026ab2/volumes" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.767662 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 06 14:14:45 crc kubenswrapper[4706]: E1206 14:14:45.768048 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1c3f32b-16af-439e-8bd4-5f0456026ab2" containerName="probe" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.768077 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1c3f32b-16af-439e-8bd4-5f0456026ab2" containerName="probe" Dec 06 14:14:45 crc kubenswrapper[4706]: E1206 14:14:45.768102 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1c3f32b-16af-439e-8bd4-5f0456026ab2" containerName="cinder-scheduler" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.768115 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1c3f32b-16af-439e-8bd4-5f0456026ab2" containerName="cinder-scheduler" Dec 06 14:14:45 crc kubenswrapper[4706]: E1206 14:14:45.768144 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e871919b-dc73-4ac8-a8fc-326200c509f9" containerName="probe" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.768155 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e871919b-dc73-4ac8-a8fc-326200c509f9" containerName="probe" Dec 06 14:14:45 crc kubenswrapper[4706]: E1206 14:14:45.768175 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e871919b-dc73-4ac8-a8fc-326200c509f9" containerName="cinder-scheduler" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.768185 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e871919b-dc73-4ac8-a8fc-326200c509f9" containerName="cinder-scheduler" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.768434 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e871919b-dc73-4ac8-a8fc-326200c509f9" containerName="cinder-scheduler" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.768465 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1c3f32b-16af-439e-8bd4-5f0456026ab2" containerName="probe" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.768595 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e871919b-dc73-4ac8-a8fc-326200c509f9" containerName="probe" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.768618 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1c3f32b-16af-439e-8bd4-5f0456026ab2" containerName="cinder-scheduler" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.769638 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.819719 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.955750 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-lib-modules\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.955853 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-dev\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.955888 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-machine-id\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.955921 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-locks-brick\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.955951 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-run\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.955981 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-config-data\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.956007 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-lib-cinder\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.956044 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-locks-cinder\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.956094 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvcg9\" (UniqueName: \"kubernetes.io/projected/c79e1377-9c84-411b-a430-671f282120cc-kube-api-access-wvcg9\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.956138 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-iscsi\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.956164 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-scripts\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.956193 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-sys\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.956221 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-config-data-custom\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:45 crc kubenswrapper[4706]: I1206 14:14:45.956354 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-nvme\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.057806 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvcg9\" (UniqueName: \"kubernetes.io/projected/c79e1377-9c84-411b-a430-671f282120cc-kube-api-access-wvcg9\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.057910 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-iscsi\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.057962 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-scripts\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.058010 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-sys\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.058051 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-config-data-custom\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.058078 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-iscsi\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.058518 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-sys\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.058731 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-nvme\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.058816 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-lib-modules\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.058871 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-nvme\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.058951 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-lib-modules\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.058882 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-dev\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.059024 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-machine-id\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.059025 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-dev\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.059073 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-locks-brick\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.059107 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-run\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.059148 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-config-data\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.059184 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-lib-cinder\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.059188 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-locks-brick\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.059189 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-run\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.059234 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-locks-cinder\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.059291 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-lib-cinder\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.059108 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-machine-id\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.059381 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-locks-cinder\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.065565 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-config-data-custom\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.066160 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-scripts\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.067888 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-config-data\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.083124 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvcg9\" (UniqueName: \"kubernetes.io/projected/c79e1377-9c84-411b-a430-671f282120cc-kube-api-access-wvcg9\") pod \"cinder-backup-1\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.090453 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:46 crc kubenswrapper[4706]: I1206 14:14:46.372223 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 06 14:14:47 crc kubenswrapper[4706]: I1206 14:14:47.317569 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"c79e1377-9c84-411b-a430-671f282120cc","Type":"ContainerStarted","Data":"cdb3f517e07f5a5fcda2d29d85423e7797e5e1961c98e0b95360b4e28982c6c7"} Dec 06 14:14:47 crc kubenswrapper[4706]: I1206 14:14:47.318213 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"c79e1377-9c84-411b-a430-671f282120cc","Type":"ContainerStarted","Data":"671210fc6c87a46a61e709de800bcd4e1a54a3bd99e161f646db1126b50db889"} Dec 06 14:14:47 crc kubenswrapper[4706]: I1206 14:14:47.318235 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"c79e1377-9c84-411b-a430-671f282120cc","Type":"ContainerStarted","Data":"6a086af64894fe60b0696f8b5268bc226e8ed18e5ab887d72312e172f6aac6f8"} Dec 06 14:14:47 crc kubenswrapper[4706]: I1206 14:14:47.360235 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-1" podStartSLOduration=2.360215778 podStartE2EDuration="2.360215778s" podCreationTimestamp="2025-12-06 14:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:14:47.356577572 +0000 UTC m=+1130.192370489" watchObservedRunningTime="2025-12-06 14:14:47.360215778 +0000 UTC m=+1130.196008665" Dec 06 14:14:50 crc kubenswrapper[4706]: I1206 14:14:50.511548 4706 scope.go:117] "RemoveContainer" containerID="eacfd1e4a7285c5d5131f1219a33c447e2090f9c230e3d482b8c62a51443c54d" Dec 06 14:14:50 crc kubenswrapper[4706]: I1206 14:14:50.512086 4706 scope.go:117] "RemoveContainer" containerID="00e2f85b4d8d2de959364092eb69e66b15126a02a2a6fedd5f3ad6a761ae3ffa" Dec 06 14:14:50 crc kubenswrapper[4706]: E1206 14:14:50.512769 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.090763 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.289352 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.418089 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.419138 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.435450 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.553681 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-iscsi\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.553764 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-sys\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.553838 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-locks-brick\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.553903 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-config-data-custom\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.553961 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-config-data\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.554114 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-nvme\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.554164 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-lib-cinder\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.554274 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-machine-id\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.554359 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-scripts\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.554476 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-run\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.554522 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-dev\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.554578 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l642m\" (UniqueName: \"kubernetes.io/projected/f7ceb31b-585b-44d8-bac3-6e729ff80572-kube-api-access-l642m\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.554678 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-locks-cinder\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.554768 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-lib-modules\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.656937 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-config-data\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.657076 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-nvme\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.657107 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-lib-cinder\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.657206 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-machine-id\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.657245 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-scripts\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.657291 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-lib-cinder\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.657346 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-run\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.657433 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-dev\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.657487 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l642m\" (UniqueName: \"kubernetes.io/projected/f7ceb31b-585b-44d8-bac3-6e729ff80572-kube-api-access-l642m\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.657533 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-locks-cinder\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.657594 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-lib-modules\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.657688 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-iscsi\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.657730 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-sys\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.657788 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-locks-brick\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.657860 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-config-data-custom\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.658161 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-lib-modules\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.658231 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-nvme\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.657388 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-run\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.658371 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-dev\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.658423 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-machine-id\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.658835 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-sys\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.659123 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-iscsi\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.659409 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-locks-cinder\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.659245 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-locks-brick\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.665379 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-scripts\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.673113 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-config-data\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.675261 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-config-data-custom\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.695609 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l642m\" (UniqueName: \"kubernetes.io/projected/f7ceb31b-585b-44d8-bac3-6e729ff80572-kube-api-access-l642m\") pod \"cinder-backup-2\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:51 crc kubenswrapper[4706]: I1206 14:14:51.758228 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:14:52 crc kubenswrapper[4706]: I1206 14:14:52.200438 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 06 14:14:52 crc kubenswrapper[4706]: I1206 14:14:52.359397 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"f7ceb31b-585b-44d8-bac3-6e729ff80572","Type":"ContainerStarted","Data":"494b2a376c7538ece355e2f7099f9b380706ee1abca53a438aa6d3cad2e0afbb"} Dec 06 14:14:53 crc kubenswrapper[4706]: I1206 14:14:53.372950 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"f7ceb31b-585b-44d8-bac3-6e729ff80572","Type":"ContainerStarted","Data":"b2733f03aa06c6cbe22681e86c722bd9c5ea0357463965053aed2391da2bf7c3"} Dec 06 14:14:53 crc kubenswrapper[4706]: I1206 14:14:53.373265 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"f7ceb31b-585b-44d8-bac3-6e729ff80572","Type":"ContainerStarted","Data":"70612e3d6f06067a2b9bcaa54a33448ce0c5a55e3bcde3c97ad90f1a7fbc06dd"} Dec 06 14:14:53 crc kubenswrapper[4706]: I1206 14:14:53.419821 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-2" podStartSLOduration=2.419801743 podStartE2EDuration="2.419801743s" podCreationTimestamp="2025-12-06 14:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:14:53.401038216 +0000 UTC m=+1136.236831103" watchObservedRunningTime="2025-12-06 14:14:53.419801743 +0000 UTC m=+1136.255594600" Dec 06 14:14:56 crc kubenswrapper[4706]: I1206 14:14:56.759259 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:15:00 crc kubenswrapper[4706]: I1206 14:15:00.160817 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm"] Dec 06 14:15:00 crc kubenswrapper[4706]: I1206 14:15:00.162605 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm" Dec 06 14:15:00 crc kubenswrapper[4706]: I1206 14:15:00.165483 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 14:15:00 crc kubenswrapper[4706]: I1206 14:15:00.165768 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 14:15:00 crc kubenswrapper[4706]: I1206 14:15:00.184597 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm"] Dec 06 14:15:00 crc kubenswrapper[4706]: I1206 14:15:00.320705 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7dadbdd-11fd-457a-9b15-8244365d52bd-secret-volume\") pod \"collect-profiles-29417175-5dqpm\" (UID: \"a7dadbdd-11fd-457a-9b15-8244365d52bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm" Dec 06 14:15:00 crc kubenswrapper[4706]: I1206 14:15:00.320822 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7dadbdd-11fd-457a-9b15-8244365d52bd-config-volume\") pod \"collect-profiles-29417175-5dqpm\" (UID: \"a7dadbdd-11fd-457a-9b15-8244365d52bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm" Dec 06 14:15:00 crc kubenswrapper[4706]: I1206 14:15:00.320993 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4pbb\" (UniqueName: \"kubernetes.io/projected/a7dadbdd-11fd-457a-9b15-8244365d52bd-kube-api-access-p4pbb\") pod \"collect-profiles-29417175-5dqpm\" (UID: \"a7dadbdd-11fd-457a-9b15-8244365d52bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm" Dec 06 14:15:00 crc kubenswrapper[4706]: I1206 14:15:00.422958 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4pbb\" (UniqueName: \"kubernetes.io/projected/a7dadbdd-11fd-457a-9b15-8244365d52bd-kube-api-access-p4pbb\") pod \"collect-profiles-29417175-5dqpm\" (UID: \"a7dadbdd-11fd-457a-9b15-8244365d52bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm" Dec 06 14:15:00 crc kubenswrapper[4706]: I1206 14:15:00.423154 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7dadbdd-11fd-457a-9b15-8244365d52bd-secret-volume\") pod \"collect-profiles-29417175-5dqpm\" (UID: \"a7dadbdd-11fd-457a-9b15-8244365d52bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm" Dec 06 14:15:00 crc kubenswrapper[4706]: I1206 14:15:00.423216 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7dadbdd-11fd-457a-9b15-8244365d52bd-config-volume\") pod \"collect-profiles-29417175-5dqpm\" (UID: \"a7dadbdd-11fd-457a-9b15-8244365d52bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm" Dec 06 14:15:00 crc kubenswrapper[4706]: I1206 14:15:00.425063 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7dadbdd-11fd-457a-9b15-8244365d52bd-config-volume\") pod \"collect-profiles-29417175-5dqpm\" (UID: \"a7dadbdd-11fd-457a-9b15-8244365d52bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm" Dec 06 14:15:00 crc kubenswrapper[4706]: I1206 14:15:00.433874 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7dadbdd-11fd-457a-9b15-8244365d52bd-secret-volume\") pod \"collect-profiles-29417175-5dqpm\" (UID: \"a7dadbdd-11fd-457a-9b15-8244365d52bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm" Dec 06 14:15:00 crc kubenswrapper[4706]: I1206 14:15:00.453731 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4pbb\" (UniqueName: \"kubernetes.io/projected/a7dadbdd-11fd-457a-9b15-8244365d52bd-kube-api-access-p4pbb\") pod \"collect-profiles-29417175-5dqpm\" (UID: \"a7dadbdd-11fd-457a-9b15-8244365d52bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm" Dec 06 14:15:00 crc kubenswrapper[4706]: I1206 14:15:00.485675 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm" Dec 06 14:15:00 crc kubenswrapper[4706]: I1206 14:15:00.721567 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm"] Dec 06 14:15:01 crc kubenswrapper[4706]: I1206 14:15:01.490483 4706 generic.go:334] "Generic (PLEG): container finished" podID="a7dadbdd-11fd-457a-9b15-8244365d52bd" containerID="bceedf3572370176143c049718a88ef7f5533d98d4eb3fde36f76ac6f2ade306" exitCode=0 Dec 06 14:15:01 crc kubenswrapper[4706]: I1206 14:15:01.490574 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm" event={"ID":"a7dadbdd-11fd-457a-9b15-8244365d52bd","Type":"ContainerDied","Data":"bceedf3572370176143c049718a88ef7f5533d98d4eb3fde36f76ac6f2ade306"} Dec 06 14:15:01 crc kubenswrapper[4706]: I1206 14:15:01.490824 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm" event={"ID":"a7dadbdd-11fd-457a-9b15-8244365d52bd","Type":"ContainerStarted","Data":"fe1cd3c19faeff3fcdd57effb932581c46b5e97761d65c92b23e808272b60432"} Dec 06 14:15:01 crc kubenswrapper[4706]: I1206 14:15:01.973415 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:15:02 crc kubenswrapper[4706]: I1206 14:15:02.847932 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm" Dec 06 14:15:02 crc kubenswrapper[4706]: I1206 14:15:02.870848 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7dadbdd-11fd-457a-9b15-8244365d52bd-config-volume\") pod \"a7dadbdd-11fd-457a-9b15-8244365d52bd\" (UID: \"a7dadbdd-11fd-457a-9b15-8244365d52bd\") " Dec 06 14:15:02 crc kubenswrapper[4706]: I1206 14:15:02.870936 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4pbb\" (UniqueName: \"kubernetes.io/projected/a7dadbdd-11fd-457a-9b15-8244365d52bd-kube-api-access-p4pbb\") pod \"a7dadbdd-11fd-457a-9b15-8244365d52bd\" (UID: \"a7dadbdd-11fd-457a-9b15-8244365d52bd\") " Dec 06 14:15:02 crc kubenswrapper[4706]: I1206 14:15:02.871025 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7dadbdd-11fd-457a-9b15-8244365d52bd-secret-volume\") pod \"a7dadbdd-11fd-457a-9b15-8244365d52bd\" (UID: \"a7dadbdd-11fd-457a-9b15-8244365d52bd\") " Dec 06 14:15:02 crc kubenswrapper[4706]: I1206 14:15:02.872161 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7dadbdd-11fd-457a-9b15-8244365d52bd-config-volume" (OuterVolumeSpecName: "config-volume") pod "a7dadbdd-11fd-457a-9b15-8244365d52bd" (UID: "a7dadbdd-11fd-457a-9b15-8244365d52bd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:15:02 crc kubenswrapper[4706]: I1206 14:15:02.880540 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7dadbdd-11fd-457a-9b15-8244365d52bd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a7dadbdd-11fd-457a-9b15-8244365d52bd" (UID: "a7dadbdd-11fd-457a-9b15-8244365d52bd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:02 crc kubenswrapper[4706]: I1206 14:15:02.881415 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7dadbdd-11fd-457a-9b15-8244365d52bd-kube-api-access-p4pbb" (OuterVolumeSpecName: "kube-api-access-p4pbb") pod "a7dadbdd-11fd-457a-9b15-8244365d52bd" (UID: "a7dadbdd-11fd-457a-9b15-8244365d52bd"). InnerVolumeSpecName "kube-api-access-p4pbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:15:02 crc kubenswrapper[4706]: I1206 14:15:02.973253 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4pbb\" (UniqueName: \"kubernetes.io/projected/a7dadbdd-11fd-457a-9b15-8244365d52bd-kube-api-access-p4pbb\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:02 crc kubenswrapper[4706]: I1206 14:15:02.973301 4706 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7dadbdd-11fd-457a-9b15-8244365d52bd-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:02 crc kubenswrapper[4706]: I1206 14:15:02.973337 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7dadbdd-11fd-457a-9b15-8244365d52bd-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:03 crc kubenswrapper[4706]: I1206 14:15:03.235229 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 06 14:15:03 crc kubenswrapper[4706]: I1206 14:15:03.235815 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-2" podUID="f7ceb31b-585b-44d8-bac3-6e729ff80572" containerName="cinder-backup" containerID="cri-o://b2733f03aa06c6cbe22681e86c722bd9c5ea0357463965053aed2391da2bf7c3" gracePeriod=30 Dec 06 14:15:03 crc kubenswrapper[4706]: I1206 14:15:03.235875 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-2" podUID="f7ceb31b-585b-44d8-bac3-6e729ff80572" containerName="probe" containerID="cri-o://70612e3d6f06067a2b9bcaa54a33448ce0c5a55e3bcde3c97ad90f1a7fbc06dd" gracePeriod=30 Dec 06 14:15:03 crc kubenswrapper[4706]: I1206 14:15:03.511977 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm" Dec 06 14:15:03 crc kubenswrapper[4706]: I1206 14:15:03.528919 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417175-5dqpm" event={"ID":"a7dadbdd-11fd-457a-9b15-8244365d52bd","Type":"ContainerDied","Data":"fe1cd3c19faeff3fcdd57effb932581c46b5e97761d65c92b23e808272b60432"} Dec 06 14:15:03 crc kubenswrapper[4706]: I1206 14:15:03.528986 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe1cd3c19faeff3fcdd57effb932581c46b5e97761d65c92b23e808272b60432" Dec 06 14:15:04 crc kubenswrapper[4706]: I1206 14:15:04.526082 4706 generic.go:334] "Generic (PLEG): container finished" podID="f7ceb31b-585b-44d8-bac3-6e729ff80572" containerID="70612e3d6f06067a2b9bcaa54a33448ce0c5a55e3bcde3c97ad90f1a7fbc06dd" exitCode=0 Dec 06 14:15:04 crc kubenswrapper[4706]: I1206 14:15:04.526144 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"f7ceb31b-585b-44d8-bac3-6e729ff80572","Type":"ContainerDied","Data":"70612e3d6f06067a2b9bcaa54a33448ce0c5a55e3bcde3c97ad90f1a7fbc06dd"} Dec 06 14:15:05 crc kubenswrapper[4706]: I1206 14:15:05.511645 4706 scope.go:117] "RemoveContainer" containerID="eacfd1e4a7285c5d5131f1219a33c447e2090f9c230e3d482b8c62a51443c54d" Dec 06 14:15:05 crc kubenswrapper[4706]: I1206 14:15:05.511948 4706 scope.go:117] "RemoveContainer" containerID="00e2f85b4d8d2de959364092eb69e66b15126a02a2a6fedd5f3ad6a761ae3ffa" Dec 06 14:15:06 crc kubenswrapper[4706]: I1206 14:15:06.546234 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerStarted","Data":"604f2eb8b124b8739f9dada14084488fa9e69b991c9482a2fd328027bb04dbc6"} Dec 06 14:15:06 crc kubenswrapper[4706]: I1206 14:15:06.546969 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerStarted","Data":"7ef4bbc8f42569aad0b46e84bb4ddecb1531e14ecef5b4eec1ad04eeb93a0b6b"} Dec 06 14:15:07 crc kubenswrapper[4706]: I1206 14:15:07.542134 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.336187 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.469791 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-locks-cinder\") pod \"f7ceb31b-585b-44d8-bac3-6e729ff80572\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.469891 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-lib-cinder\") pod \"f7ceb31b-585b-44d8-bac3-6e729ff80572\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.469917 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "f7ceb31b-585b-44d8-bac3-6e729ff80572" (UID: "f7ceb31b-585b-44d8-bac3-6e729ff80572"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.469995 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-sys\") pod \"f7ceb31b-585b-44d8-bac3-6e729ff80572\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.470007 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "f7ceb31b-585b-44d8-bac3-6e729ff80572" (UID: "f7ceb31b-585b-44d8-bac3-6e729ff80572"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.470060 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-locks-brick\") pod \"f7ceb31b-585b-44d8-bac3-6e729ff80572\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.470109 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-scripts\") pod \"f7ceb31b-585b-44d8-bac3-6e729ff80572\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.470116 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-sys" (OuterVolumeSpecName: "sys") pod "f7ceb31b-585b-44d8-bac3-6e729ff80572" (UID: "f7ceb31b-585b-44d8-bac3-6e729ff80572"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.470148 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "f7ceb31b-585b-44d8-bac3-6e729ff80572" (UID: "f7ceb31b-585b-44d8-bac3-6e729ff80572"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.470224 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-run\") pod \"f7ceb31b-585b-44d8-bac3-6e729ff80572\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.470352 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-run" (OuterVolumeSpecName: "run") pod "f7ceb31b-585b-44d8-bac3-6e729ff80572" (UID: "f7ceb31b-585b-44d8-bac3-6e729ff80572"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.470383 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-iscsi\") pod \"f7ceb31b-585b-44d8-bac3-6e729ff80572\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.470507 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "f7ceb31b-585b-44d8-bac3-6e729ff80572" (UID: "f7ceb31b-585b-44d8-bac3-6e729ff80572"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.470606 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-config-data-custom\") pod \"f7ceb31b-585b-44d8-bac3-6e729ff80572\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.471691 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-dev\") pod \"f7ceb31b-585b-44d8-bac3-6e729ff80572\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.471741 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-config-data\") pod \"f7ceb31b-585b-44d8-bac3-6e729ff80572\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.471770 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-nvme\") pod \"f7ceb31b-585b-44d8-bac3-6e729ff80572\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.471817 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-dev" (OuterVolumeSpecName: "dev") pod "f7ceb31b-585b-44d8-bac3-6e729ff80572" (UID: "f7ceb31b-585b-44d8-bac3-6e729ff80572"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.471822 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-machine-id\") pod \"f7ceb31b-585b-44d8-bac3-6e729ff80572\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.471872 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f7ceb31b-585b-44d8-bac3-6e729ff80572" (UID: "f7ceb31b-585b-44d8-bac3-6e729ff80572"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.471910 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l642m\" (UniqueName: \"kubernetes.io/projected/f7ceb31b-585b-44d8-bac3-6e729ff80572-kube-api-access-l642m\") pod \"f7ceb31b-585b-44d8-bac3-6e729ff80572\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.471971 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-lib-modules\") pod \"f7ceb31b-585b-44d8-bac3-6e729ff80572\" (UID: \"f7ceb31b-585b-44d8-bac3-6e729ff80572\") " Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.471984 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "f7ceb31b-585b-44d8-bac3-6e729ff80572" (UID: "f7ceb31b-585b-44d8-bac3-6e729ff80572"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.472177 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "f7ceb31b-585b-44d8-bac3-6e729ff80572" (UID: "f7ceb31b-585b-44d8-bac3-6e729ff80572"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.472683 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.472720 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-dev\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.472741 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.472760 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.472777 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.472795 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.472813 4706 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.472830 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-sys\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.472847 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.472864 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f7ceb31b-585b-44d8-bac3-6e729ff80572-run\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.478968 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7ceb31b-585b-44d8-bac3-6e729ff80572-kube-api-access-l642m" (OuterVolumeSpecName: "kube-api-access-l642m") pod "f7ceb31b-585b-44d8-bac3-6e729ff80572" (UID: "f7ceb31b-585b-44d8-bac3-6e729ff80572"). InnerVolumeSpecName "kube-api-access-l642m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.483682 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-scripts" (OuterVolumeSpecName: "scripts") pod "f7ceb31b-585b-44d8-bac3-6e729ff80572" (UID: "f7ceb31b-585b-44d8-bac3-6e729ff80572"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.495101 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f7ceb31b-585b-44d8-bac3-6e729ff80572" (UID: "f7ceb31b-585b-44d8-bac3-6e729ff80572"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.573857 4706 generic.go:334] "Generic (PLEG): container finished" podID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerID="604f2eb8b124b8739f9dada14084488fa9e69b991c9482a2fd328027bb04dbc6" exitCode=1 Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.573999 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerDied","Data":"604f2eb8b124b8739f9dada14084488fa9e69b991c9482a2fd328027bb04dbc6"} Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.574364 4706 scope.go:117] "RemoveContainer" containerID="00e2f85b4d8d2de959364092eb69e66b15126a02a2a6fedd5f3ad6a761ae3ffa" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.574673 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.574720 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.574743 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l642m\" (UniqueName: \"kubernetes.io/projected/f7ceb31b-585b-44d8-bac3-6e729ff80572-kube-api-access-l642m\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.576368 4706 scope.go:117] "RemoveContainer" containerID="604f2eb8b124b8739f9dada14084488fa9e69b991c9482a2fd328027bb04dbc6" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.576631 4706 generic.go:334] "Generic (PLEG): container finished" podID="f7ceb31b-585b-44d8-bac3-6e729ff80572" containerID="b2733f03aa06c6cbe22681e86c722bd9c5ea0357463965053aed2391da2bf7c3" exitCode=0 Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.576666 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"f7ceb31b-585b-44d8-bac3-6e729ff80572","Type":"ContainerDied","Data":"b2733f03aa06c6cbe22681e86c722bd9c5ea0357463965053aed2391da2bf7c3"} Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.576690 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"f7ceb31b-585b-44d8-bac3-6e729ff80572","Type":"ContainerDied","Data":"494b2a376c7538ece355e2f7099f9b380706ee1abca53a438aa6d3cad2e0afbb"} Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.576807 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-2" Dec 06 14:15:08 crc kubenswrapper[4706]: E1206 14:15:08.577051 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.585348 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-config-data" (OuterVolumeSpecName: "config-data") pod "f7ceb31b-585b-44d8-bac3-6e729ff80572" (UID: "f7ceb31b-585b-44d8-bac3-6e729ff80572"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.665896 4706 scope.go:117] "RemoveContainer" containerID="70612e3d6f06067a2b9bcaa54a33448ce0c5a55e3bcde3c97ad90f1a7fbc06dd" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.677071 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ceb31b-585b-44d8-bac3-6e729ff80572-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.682027 4706 scope.go:117] "RemoveContainer" containerID="b2733f03aa06c6cbe22681e86c722bd9c5ea0357463965053aed2391da2bf7c3" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.701982 4706 scope.go:117] "RemoveContainer" containerID="70612e3d6f06067a2b9bcaa54a33448ce0c5a55e3bcde3c97ad90f1a7fbc06dd" Dec 06 14:15:08 crc kubenswrapper[4706]: E1206 14:15:08.702518 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70612e3d6f06067a2b9bcaa54a33448ce0c5a55e3bcde3c97ad90f1a7fbc06dd\": container with ID starting with 70612e3d6f06067a2b9bcaa54a33448ce0c5a55e3bcde3c97ad90f1a7fbc06dd not found: ID does not exist" containerID="70612e3d6f06067a2b9bcaa54a33448ce0c5a55e3bcde3c97ad90f1a7fbc06dd" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.702568 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70612e3d6f06067a2b9bcaa54a33448ce0c5a55e3bcde3c97ad90f1a7fbc06dd"} err="failed to get container status \"70612e3d6f06067a2b9bcaa54a33448ce0c5a55e3bcde3c97ad90f1a7fbc06dd\": rpc error: code = NotFound desc = could not find container \"70612e3d6f06067a2b9bcaa54a33448ce0c5a55e3bcde3c97ad90f1a7fbc06dd\": container with ID starting with 70612e3d6f06067a2b9bcaa54a33448ce0c5a55e3bcde3c97ad90f1a7fbc06dd not found: ID does not exist" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.702599 4706 scope.go:117] "RemoveContainer" containerID="b2733f03aa06c6cbe22681e86c722bd9c5ea0357463965053aed2391da2bf7c3" Dec 06 14:15:08 crc kubenswrapper[4706]: E1206 14:15:08.702965 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2733f03aa06c6cbe22681e86c722bd9c5ea0357463965053aed2391da2bf7c3\": container with ID starting with b2733f03aa06c6cbe22681e86c722bd9c5ea0357463965053aed2391da2bf7c3 not found: ID does not exist" containerID="b2733f03aa06c6cbe22681e86c722bd9c5ea0357463965053aed2391da2bf7c3" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.703009 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2733f03aa06c6cbe22681e86c722bd9c5ea0357463965053aed2391da2bf7c3"} err="failed to get container status \"b2733f03aa06c6cbe22681e86c722bd9c5ea0357463965053aed2391da2bf7c3\": rpc error: code = NotFound desc = could not find container \"b2733f03aa06c6cbe22681e86c722bd9c5ea0357463965053aed2391da2bf7c3\": container with ID starting with b2733f03aa06c6cbe22681e86c722bd9c5ea0357463965053aed2391da2bf7c3 not found: ID does not exist" Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.931299 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.945634 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.958071 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.958468 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-1" podUID="c79e1377-9c84-411b-a430-671f282120cc" containerName="cinder-backup" containerID="cri-o://671210fc6c87a46a61e709de800bcd4e1a54a3bd99e161f646db1126b50db889" gracePeriod=30 Dec 06 14:15:08 crc kubenswrapper[4706]: I1206 14:15:08.958633 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-1" podUID="c79e1377-9c84-411b-a430-671f282120cc" containerName="probe" containerID="cri-o://cdb3f517e07f5a5fcda2d29d85423e7797e5e1961c98e0b95360b4e28982c6c7" gracePeriod=30 Dec 06 14:15:09 crc kubenswrapper[4706]: I1206 14:15:09.527799 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7ceb31b-585b-44d8-bac3-6e729ff80572" path="/var/lib/kubelet/pods/f7ceb31b-585b-44d8-bac3-6e729ff80572/volumes" Dec 06 14:15:09 crc kubenswrapper[4706]: I1206 14:15:09.588773 4706 generic.go:334] "Generic (PLEG): container finished" podID="c79e1377-9c84-411b-a430-671f282120cc" containerID="cdb3f517e07f5a5fcda2d29d85423e7797e5e1961c98e0b95360b4e28982c6c7" exitCode=0 Dec 06 14:15:09 crc kubenswrapper[4706]: I1206 14:15:09.588826 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"c79e1377-9c84-411b-a430-671f282120cc","Type":"ContainerDied","Data":"cdb3f517e07f5a5fcda2d29d85423e7797e5e1961c98e0b95360b4e28982c6c7"} Dec 06 14:15:09 crc kubenswrapper[4706]: I1206 14:15:09.592599 4706 generic.go:334] "Generic (PLEG): container finished" podID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerID="7ef4bbc8f42569aad0b46e84bb4ddecb1531e14ecef5b4eec1ad04eeb93a0b6b" exitCode=1 Dec 06 14:15:09 crc kubenswrapper[4706]: I1206 14:15:09.592690 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerDied","Data":"7ef4bbc8f42569aad0b46e84bb4ddecb1531e14ecef5b4eec1ad04eeb93a0b6b"} Dec 06 14:15:09 crc kubenswrapper[4706]: I1206 14:15:09.592745 4706 scope.go:117] "RemoveContainer" containerID="eacfd1e4a7285c5d5131f1219a33c447e2090f9c230e3d482b8c62a51443c54d" Dec 06 14:15:09 crc kubenswrapper[4706]: I1206 14:15:09.593367 4706 scope.go:117] "RemoveContainer" containerID="7ef4bbc8f42569aad0b46e84bb4ddecb1531e14ecef5b4eec1ad04eeb93a0b6b" Dec 06 14:15:09 crc kubenswrapper[4706]: I1206 14:15:09.593413 4706 scope.go:117] "RemoveContainer" containerID="604f2eb8b124b8739f9dada14084488fa9e69b991c9482a2fd328027bb04dbc6" Dec 06 14:15:09 crc kubenswrapper[4706]: E1206 14:15:09.593818 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" Dec 06 14:15:11 crc kubenswrapper[4706]: I1206 14:15:11.542047 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:11 crc kubenswrapper[4706]: I1206 14:15:11.543894 4706 scope.go:117] "RemoveContainer" containerID="7ef4bbc8f42569aad0b46e84bb4ddecb1531e14ecef5b4eec1ad04eeb93a0b6b" Dec 06 14:15:11 crc kubenswrapper[4706]: I1206 14:15:11.544021 4706 scope.go:117] "RemoveContainer" containerID="604f2eb8b124b8739f9dada14084488fa9e69b991c9482a2fd328027bb04dbc6" Dec 06 14:15:11 crc kubenswrapper[4706]: E1206 14:15:11.544481 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" Dec 06 14:15:12 crc kubenswrapper[4706]: I1206 14:15:12.541370 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:12 crc kubenswrapper[4706]: I1206 14:15:12.542447 4706 scope.go:117] "RemoveContainer" containerID="7ef4bbc8f42569aad0b46e84bb4ddecb1531e14ecef5b4eec1ad04eeb93a0b6b" Dec 06 14:15:12 crc kubenswrapper[4706]: I1206 14:15:12.542480 4706 scope.go:117] "RemoveContainer" containerID="604f2eb8b124b8739f9dada14084488fa9e69b991c9482a2fd328027bb04dbc6" Dec 06 14:15:12 crc kubenswrapper[4706]: E1206 14:15:12.542861 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" Dec 06 14:15:13 crc kubenswrapper[4706]: E1206 14:15:13.504718 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc79e1377_9c84_411b_a430_671f282120cc.slice/crio-671210fc6c87a46a61e709de800bcd4e1a54a3bd99e161f646db1126b50db889.scope\": RecentStats: unable to find data in memory cache]" Dec 06 14:15:13 crc kubenswrapper[4706]: I1206 14:15:13.656044 4706 generic.go:334] "Generic (PLEG): container finished" podID="c79e1377-9c84-411b-a430-671f282120cc" containerID="671210fc6c87a46a61e709de800bcd4e1a54a3bd99e161f646db1126b50db889" exitCode=0 Dec 06 14:15:13 crc kubenswrapper[4706]: I1206 14:15:13.656089 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"c79e1377-9c84-411b-a430-671f282120cc","Type":"ContainerDied","Data":"671210fc6c87a46a61e709de800bcd4e1a54a3bd99e161f646db1126b50db889"} Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.458757 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.574818 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-machine-id\") pod \"c79e1377-9c84-411b-a430-671f282120cc\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.574875 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-sys\") pod \"c79e1377-9c84-411b-a430-671f282120cc\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.574933 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-scripts\") pod \"c79e1377-9c84-411b-a430-671f282120cc\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.574958 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-config-data\") pod \"c79e1377-9c84-411b-a430-671f282120cc\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575009 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-locks-cinder\") pod \"c79e1377-9c84-411b-a430-671f282120cc\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575048 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-nvme\") pod \"c79e1377-9c84-411b-a430-671f282120cc\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575072 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvcg9\" (UniqueName: \"kubernetes.io/projected/c79e1377-9c84-411b-a430-671f282120cc-kube-api-access-wvcg9\") pod \"c79e1377-9c84-411b-a430-671f282120cc\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575092 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-iscsi\") pod \"c79e1377-9c84-411b-a430-671f282120cc\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575128 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-lib-cinder\") pod \"c79e1377-9c84-411b-a430-671f282120cc\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575149 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-dev\") pod \"c79e1377-9c84-411b-a430-671f282120cc\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575178 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-lib-modules\") pod \"c79e1377-9c84-411b-a430-671f282120cc\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575205 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-locks-brick\") pod \"c79e1377-9c84-411b-a430-671f282120cc\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575228 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-run\") pod \"c79e1377-9c84-411b-a430-671f282120cc\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575263 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-config-data-custom\") pod \"c79e1377-9c84-411b-a430-671f282120cc\" (UID: \"c79e1377-9c84-411b-a430-671f282120cc\") " Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575543 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "c79e1377-9c84-411b-a430-671f282120cc" (UID: "c79e1377-9c84-411b-a430-671f282120cc"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575579 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c79e1377-9c84-411b-a430-671f282120cc" (UID: "c79e1377-9c84-411b-a430-671f282120cc"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575591 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-sys" (OuterVolumeSpecName: "sys") pod "c79e1377-9c84-411b-a430-671f282120cc" (UID: "c79e1377-9c84-411b-a430-671f282120cc"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575617 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "c79e1377-9c84-411b-a430-671f282120cc" (UID: "c79e1377-9c84-411b-a430-671f282120cc"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575634 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "c79e1377-9c84-411b-a430-671f282120cc" (UID: "c79e1377-9c84-411b-a430-671f282120cc"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575627 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "c79e1377-9c84-411b-a430-671f282120cc" (UID: "c79e1377-9c84-411b-a430-671f282120cc"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575675 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-run" (OuterVolumeSpecName: "run") pod "c79e1377-9c84-411b-a430-671f282120cc" (UID: "c79e1377-9c84-411b-a430-671f282120cc"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575652 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "c79e1377-9c84-411b-a430-671f282120cc" (UID: "c79e1377-9c84-411b-a430-671f282120cc"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575651 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-dev" (OuterVolumeSpecName: "dev") pod "c79e1377-9c84-411b-a430-671f282120cc" (UID: "c79e1377-9c84-411b-a430-671f282120cc"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.575660 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "c79e1377-9c84-411b-a430-671f282120cc" (UID: "c79e1377-9c84-411b-a430-671f282120cc"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.579542 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.579597 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.581465 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-scripts" (OuterVolumeSpecName: "scripts") pod "c79e1377-9c84-411b-a430-671f282120cc" (UID: "c79e1377-9c84-411b-a430-671f282120cc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.585504 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c79e1377-9c84-411b-a430-671f282120cc-kube-api-access-wvcg9" (OuterVolumeSpecName: "kube-api-access-wvcg9") pod "c79e1377-9c84-411b-a430-671f282120cc" (UID: "c79e1377-9c84-411b-a430-671f282120cc"). InnerVolumeSpecName "kube-api-access-wvcg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.601412 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c79e1377-9c84-411b-a430-671f282120cc" (UID: "c79e1377-9c84-411b-a430-671f282120cc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.646131 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-config-data" (OuterVolumeSpecName: "config-data") pod "c79e1377-9c84-411b-a430-671f282120cc" (UID: "c79e1377-9c84-411b-a430-671f282120cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.667407 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"c79e1377-9c84-411b-a430-671f282120cc","Type":"ContainerDied","Data":"6a086af64894fe60b0696f8b5268bc226e8ed18e5ab887d72312e172f6aac6f8"} Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.667451 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-1" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.667457 4706 scope.go:117] "RemoveContainer" containerID="cdb3f517e07f5a5fcda2d29d85423e7797e5e1961c98e0b95360b4e28982c6c7" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.676793 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.676847 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-sys\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.676866 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.676883 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.676900 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.676917 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.676933 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvcg9\" (UniqueName: \"kubernetes.io/projected/c79e1377-9c84-411b-a430-671f282120cc-kube-api-access-wvcg9\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.676957 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.676980 4706 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.677002 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-dev\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.677026 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.677043 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.677061 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c79e1377-9c84-411b-a430-671f282120cc-run\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.677078 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c79e1377-9c84-411b-a430-671f282120cc-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.697559 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.698430 4706 scope.go:117] "RemoveContainer" containerID="671210fc6c87a46a61e709de800bcd4e1a54a3bd99e161f646db1126b50db889" Dec 06 14:15:14 crc kubenswrapper[4706]: I1206 14:15:14.702458 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 06 14:15:15 crc kubenswrapper[4706]: I1206 14:15:15.527653 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c79e1377-9c84-411b-a430-671f282120cc" path="/var/lib/kubelet/pods/c79e1377-9c84-411b-a430-671f282120cc/volumes" Dec 06 14:15:15 crc kubenswrapper[4706]: I1206 14:15:15.768171 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:15:15 crc kubenswrapper[4706]: I1206 14:15:15.773716 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="f43169ab-3a35-4d23-8758-9e63360c9844" containerName="cinder-api-log" containerID="cri-o://ebc70353af85aec8e06892c24ab6c1d7b5df3da8ea1058d9b32b9b0ae3184e27" gracePeriod=30 Dec 06 14:15:15 crc kubenswrapper[4706]: I1206 14:15:15.773798 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="f43169ab-3a35-4d23-8758-9e63360c9844" containerName="cinder-api" containerID="cri-o://2f18ca8696222578e2b0d0a3d4517e4945077dc4ebfe9588fab77f913b5e5c27" gracePeriod=30 Dec 06 14:15:16 crc kubenswrapper[4706]: I1206 14:15:16.692088 4706 generic.go:334] "Generic (PLEG): container finished" podID="f43169ab-3a35-4d23-8758-9e63360c9844" containerID="ebc70353af85aec8e06892c24ab6c1d7b5df3da8ea1058d9b32b9b0ae3184e27" exitCode=143 Dec 06 14:15:16 crc kubenswrapper[4706]: I1206 14:15:16.692164 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"f43169ab-3a35-4d23-8758-9e63360c9844","Type":"ContainerDied","Data":"ebc70353af85aec8e06892c24ab6c1d7b5df3da8ea1058d9b32b9b0ae3184e27"} Dec 06 14:15:18 crc kubenswrapper[4706]: I1206 14:15:18.926686 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-0" podUID="f43169ab-3a35-4d23-8758-9e63360c9844" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.91:8776/healthcheck\": read tcp 10.217.0.2:40498->10.217.0.91:8776: read: connection reset by peer" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.378739 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.454015 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-scripts\") pod \"f43169ab-3a35-4d23-8758-9e63360c9844\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.454072 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2v25p\" (UniqueName: \"kubernetes.io/projected/f43169ab-3a35-4d23-8758-9e63360c9844-kube-api-access-2v25p\") pod \"f43169ab-3a35-4d23-8758-9e63360c9844\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.454095 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f43169ab-3a35-4d23-8758-9e63360c9844-etc-machine-id\") pod \"f43169ab-3a35-4d23-8758-9e63360c9844\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.454140 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-config-data\") pod \"f43169ab-3a35-4d23-8758-9e63360c9844\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.454297 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-config-data-custom\") pod \"f43169ab-3a35-4d23-8758-9e63360c9844\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.454349 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f43169ab-3a35-4d23-8758-9e63360c9844-logs\") pod \"f43169ab-3a35-4d23-8758-9e63360c9844\" (UID: \"f43169ab-3a35-4d23-8758-9e63360c9844\") " Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.455354 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f43169ab-3a35-4d23-8758-9e63360c9844-logs" (OuterVolumeSpecName: "logs") pod "f43169ab-3a35-4d23-8758-9e63360c9844" (UID: "f43169ab-3a35-4d23-8758-9e63360c9844"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.459452 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f43169ab-3a35-4d23-8758-9e63360c9844-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f43169ab-3a35-4d23-8758-9e63360c9844" (UID: "f43169ab-3a35-4d23-8758-9e63360c9844"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.463929 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-scripts" (OuterVolumeSpecName: "scripts") pod "f43169ab-3a35-4d23-8758-9e63360c9844" (UID: "f43169ab-3a35-4d23-8758-9e63360c9844"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.464721 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f43169ab-3a35-4d23-8758-9e63360c9844" (UID: "f43169ab-3a35-4d23-8758-9e63360c9844"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.469507 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f43169ab-3a35-4d23-8758-9e63360c9844-kube-api-access-2v25p" (OuterVolumeSpecName: "kube-api-access-2v25p") pod "f43169ab-3a35-4d23-8758-9e63360c9844" (UID: "f43169ab-3a35-4d23-8758-9e63360c9844"). InnerVolumeSpecName "kube-api-access-2v25p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.496919 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-config-data" (OuterVolumeSpecName: "config-data") pod "f43169ab-3a35-4d23-8758-9e63360c9844" (UID: "f43169ab-3a35-4d23-8758-9e63360c9844"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.555799 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.555857 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f43169ab-3a35-4d23-8758-9e63360c9844-logs\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.555878 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.555898 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2v25p\" (UniqueName: \"kubernetes.io/projected/f43169ab-3a35-4d23-8758-9e63360c9844-kube-api-access-2v25p\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.555915 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f43169ab-3a35-4d23-8758-9e63360c9844-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.555928 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f43169ab-3a35-4d23-8758-9e63360c9844-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.725060 4706 generic.go:334] "Generic (PLEG): container finished" podID="f43169ab-3a35-4d23-8758-9e63360c9844" containerID="2f18ca8696222578e2b0d0a3d4517e4945077dc4ebfe9588fab77f913b5e5c27" exitCode=0 Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.725111 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"f43169ab-3a35-4d23-8758-9e63360c9844","Type":"ContainerDied","Data":"2f18ca8696222578e2b0d0a3d4517e4945077dc4ebfe9588fab77f913b5e5c27"} Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.725140 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.725166 4706 scope.go:117] "RemoveContainer" containerID="2f18ca8696222578e2b0d0a3d4517e4945077dc4ebfe9588fab77f913b5e5c27" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.725149 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"f43169ab-3a35-4d23-8758-9e63360c9844","Type":"ContainerDied","Data":"71f84d736187d94e242d3c45f580ef46ebe9c369ad86351e20d1959dda440b57"} Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.750962 4706 scope.go:117] "RemoveContainer" containerID="ebc70353af85aec8e06892c24ab6c1d7b5df3da8ea1058d9b32b9b0ae3184e27" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.766519 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.772622 4706 scope.go:117] "RemoveContainer" containerID="2f18ca8696222578e2b0d0a3d4517e4945077dc4ebfe9588fab77f913b5e5c27" Dec 06 14:15:19 crc kubenswrapper[4706]: E1206 14:15:19.773250 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f18ca8696222578e2b0d0a3d4517e4945077dc4ebfe9588fab77f913b5e5c27\": container with ID starting with 2f18ca8696222578e2b0d0a3d4517e4945077dc4ebfe9588fab77f913b5e5c27 not found: ID does not exist" containerID="2f18ca8696222578e2b0d0a3d4517e4945077dc4ebfe9588fab77f913b5e5c27" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.773360 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f18ca8696222578e2b0d0a3d4517e4945077dc4ebfe9588fab77f913b5e5c27"} err="failed to get container status \"2f18ca8696222578e2b0d0a3d4517e4945077dc4ebfe9588fab77f913b5e5c27\": rpc error: code = NotFound desc = could not find container \"2f18ca8696222578e2b0d0a3d4517e4945077dc4ebfe9588fab77f913b5e5c27\": container with ID starting with 2f18ca8696222578e2b0d0a3d4517e4945077dc4ebfe9588fab77f913b5e5c27 not found: ID does not exist" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.773402 4706 scope.go:117] "RemoveContainer" containerID="ebc70353af85aec8e06892c24ab6c1d7b5df3da8ea1058d9b32b9b0ae3184e27" Dec 06 14:15:19 crc kubenswrapper[4706]: E1206 14:15:19.774168 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebc70353af85aec8e06892c24ab6c1d7b5df3da8ea1058d9b32b9b0ae3184e27\": container with ID starting with ebc70353af85aec8e06892c24ab6c1d7b5df3da8ea1058d9b32b9b0ae3184e27 not found: ID does not exist" containerID="ebc70353af85aec8e06892c24ab6c1d7b5df3da8ea1058d9b32b9b0ae3184e27" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.774250 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebc70353af85aec8e06892c24ab6c1d7b5df3da8ea1058d9b32b9b0ae3184e27"} err="failed to get container status \"ebc70353af85aec8e06892c24ab6c1d7b5df3da8ea1058d9b32b9b0ae3184e27\": rpc error: code = NotFound desc = could not find container \"ebc70353af85aec8e06892c24ab6c1d7b5df3da8ea1058d9b32b9b0ae3184e27\": container with ID starting with ebc70353af85aec8e06892c24ab6c1d7b5df3da8ea1058d9b32b9b0ae3184e27 not found: ID does not exist" Dec 06 14:15:19 crc kubenswrapper[4706]: I1206 14:15:19.779143 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.065781 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:15:20 crc kubenswrapper[4706]: E1206 14:15:20.067224 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ceb31b-585b-44d8-bac3-6e729ff80572" containerName="probe" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.067409 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ceb31b-585b-44d8-bac3-6e729ff80572" containerName="probe" Dec 06 14:15:20 crc kubenswrapper[4706]: E1206 14:15:20.067570 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f43169ab-3a35-4d23-8758-9e63360c9844" containerName="cinder-api" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.067696 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f43169ab-3a35-4d23-8758-9e63360c9844" containerName="cinder-api" Dec 06 14:15:20 crc kubenswrapper[4706]: E1206 14:15:20.067813 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c79e1377-9c84-411b-a430-671f282120cc" containerName="cinder-backup" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.067919 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c79e1377-9c84-411b-a430-671f282120cc" containerName="cinder-backup" Dec 06 14:15:20 crc kubenswrapper[4706]: E1206 14:15:20.068060 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7dadbdd-11fd-457a-9b15-8244365d52bd" containerName="collect-profiles" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.068179 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7dadbdd-11fd-457a-9b15-8244365d52bd" containerName="collect-profiles" Dec 06 14:15:20 crc kubenswrapper[4706]: E1206 14:15:20.068338 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c79e1377-9c84-411b-a430-671f282120cc" containerName="probe" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.068453 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c79e1377-9c84-411b-a430-671f282120cc" containerName="probe" Dec 06 14:15:20 crc kubenswrapper[4706]: E1206 14:15:20.068564 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f43169ab-3a35-4d23-8758-9e63360c9844" containerName="cinder-api-log" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.068670 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f43169ab-3a35-4d23-8758-9e63360c9844" containerName="cinder-api-log" Dec 06 14:15:20 crc kubenswrapper[4706]: E1206 14:15:20.068807 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ceb31b-585b-44d8-bac3-6e729ff80572" containerName="cinder-backup" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.068925 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ceb31b-585b-44d8-bac3-6e729ff80572" containerName="cinder-backup" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.069346 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f43169ab-3a35-4d23-8758-9e63360c9844" containerName="cinder-api-log" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.069537 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7dadbdd-11fd-457a-9b15-8244365d52bd" containerName="collect-profiles" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.069666 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f43169ab-3a35-4d23-8758-9e63360c9844" containerName="cinder-api" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.069796 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c79e1377-9c84-411b-a430-671f282120cc" containerName="probe" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.069913 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c79e1377-9c84-411b-a430-671f282120cc" containerName="cinder-backup" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.070073 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7ceb31b-585b-44d8-bac3-6e729ff80572" containerName="cinder-backup" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.070217 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7ceb31b-585b-44d8-bac3-6e729ff80572" containerName="probe" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.071477 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.075091 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.088511 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-api-config-data" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.115186 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.149370 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.150447 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.164134 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34735cc4-f322-45b7-b2cd-7dd16109952f-logs\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.164360 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5hhr\" (UniqueName: \"kubernetes.io/projected/34735cc4-f322-45b7-b2cd-7dd16109952f-kube-api-access-d5hhr\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.164487 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-config-data\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.164624 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn94x\" (UniqueName: \"kubernetes.io/projected/5f976297-d9ed-4fcf-a715-22d963cacb04-kube-api-access-dn94x\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.164756 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-config-data\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.164857 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-scripts\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.165132 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5f976297-d9ed-4fcf-a715-22d963cacb04-etc-machine-id\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.165217 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-scripts\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.165358 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-config-data-custom\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.165461 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f976297-d9ed-4fcf-a715-22d963cacb04-logs\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.165554 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-config-data-custom\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.165670 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34735cc4-f322-45b7-b2cd-7dd16109952f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.175980 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.184867 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.194382 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267421 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34735cc4-f322-45b7-b2cd-7dd16109952f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267488 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34735cc4-f322-45b7-b2cd-7dd16109952f-logs\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267527 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/572aefd5-0c7c-4609-9852-582b7b3071c6-etc-machine-id\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267547 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5hhr\" (UniqueName: \"kubernetes.io/projected/34735cc4-f322-45b7-b2cd-7dd16109952f-kube-api-access-d5hhr\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267563 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/572aefd5-0c7c-4609-9852-582b7b3071c6-logs\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267579 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-config-data-custom\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267601 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-config-data\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267618 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-scripts\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267636 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn94x\" (UniqueName: \"kubernetes.io/projected/5f976297-d9ed-4fcf-a715-22d963cacb04-kube-api-access-dn94x\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267660 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-config-data\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267680 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-scripts\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267693 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5f976297-d9ed-4fcf-a715-22d963cacb04-etc-machine-id\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267706 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-scripts\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267719 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-config-data\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267752 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-config-data-custom\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267779 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f976297-d9ed-4fcf-a715-22d963cacb04-logs\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267799 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-config-data-custom\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.267819 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mddgp\" (UniqueName: \"kubernetes.io/projected/572aefd5-0c7c-4609-9852-582b7b3071c6-kube-api-access-mddgp\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.268690 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34735cc4-f322-45b7-b2cd-7dd16109952f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.269520 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34735cc4-f322-45b7-b2cd-7dd16109952f-logs\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.269941 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f976297-d9ed-4fcf-a715-22d963cacb04-logs\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.270094 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5f976297-d9ed-4fcf-a715-22d963cacb04-etc-machine-id\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.274017 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-config-data-custom\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.274046 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-scripts\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.275017 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-config-data\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.277015 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-scripts\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.277441 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-config-data\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.284698 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn94x\" (UniqueName: \"kubernetes.io/projected/5f976297-d9ed-4fcf-a715-22d963cacb04-kube-api-access-dn94x\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.285298 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-config-data-custom\") pod \"cinder-api-2\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.288931 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5hhr\" (UniqueName: \"kubernetes.io/projected/34735cc4-f322-45b7-b2cd-7dd16109952f-kube-api-access-d5hhr\") pod \"cinder-api-0\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.370971 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mddgp\" (UniqueName: \"kubernetes.io/projected/572aefd5-0c7c-4609-9852-582b7b3071c6-kube-api-access-mddgp\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.371269 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/572aefd5-0c7c-4609-9852-582b7b3071c6-etc-machine-id\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.371380 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/572aefd5-0c7c-4609-9852-582b7b3071c6-logs\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.371455 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-config-data-custom\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.371547 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-scripts\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.371718 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/572aefd5-0c7c-4609-9852-582b7b3071c6-logs\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.371719 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-config-data\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.371386 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/572aefd5-0c7c-4609-9852-582b7b3071c6-etc-machine-id\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.375692 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-scripts\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.376207 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-config-data-custom\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.379867 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-config-data\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.385954 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mddgp\" (UniqueName: \"kubernetes.io/projected/572aefd5-0c7c-4609-9852-582b7b3071c6-kube-api-access-mddgp\") pod \"cinder-api-1\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.399181 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.466660 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.480922 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.736605 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.769886 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 06 14:15:20 crc kubenswrapper[4706]: W1206 14:15:20.773098 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod572aefd5_0c7c_4609_9852_582b7b3071c6.slice/crio-7e6636685549b062083e36e81ee5b76ebee7fd2f5978499e9d564b2d42c89bbb WatchSource:0}: Error finding container 7e6636685549b062083e36e81ee5b76ebee7fd2f5978499e9d564b2d42c89bbb: Status 404 returned error can't find the container with id 7e6636685549b062083e36e81ee5b76ebee7fd2f5978499e9d564b2d42c89bbb Dec 06 14:15:20 crc kubenswrapper[4706]: I1206 14:15:20.860696 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:15:20 crc kubenswrapper[4706]: W1206 14:15:20.865881 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34735cc4_f322_45b7_b2cd_7dd16109952f.slice/crio-92f5f6a5c272d2e4efca06f4de223fc48cdde09a19888a0c8b9979696c026c4b WatchSource:0}: Error finding container 92f5f6a5c272d2e4efca06f4de223fc48cdde09a19888a0c8b9979696c026c4b: Status 404 returned error can't find the container with id 92f5f6a5c272d2e4efca06f4de223fc48cdde09a19888a0c8b9979696c026c4b Dec 06 14:15:21 crc kubenswrapper[4706]: I1206 14:15:21.522851 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f43169ab-3a35-4d23-8758-9e63360c9844" path="/var/lib/kubelet/pods/f43169ab-3a35-4d23-8758-9e63360c9844/volumes" Dec 06 14:15:21 crc kubenswrapper[4706]: I1206 14:15:21.749224 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"572aefd5-0c7c-4609-9852-582b7b3071c6","Type":"ContainerStarted","Data":"5c4233562a3c60346725dac6f7ca868a58d2aca379fdcae21a5964b96c5b7571"} Dec 06 14:15:21 crc kubenswrapper[4706]: I1206 14:15:21.749270 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"572aefd5-0c7c-4609-9852-582b7b3071c6","Type":"ContainerStarted","Data":"7e6636685549b062083e36e81ee5b76ebee7fd2f5978499e9d564b2d42c89bbb"} Dec 06 14:15:21 crc kubenswrapper[4706]: I1206 14:15:21.751837 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"5f976297-d9ed-4fcf-a715-22d963cacb04","Type":"ContainerStarted","Data":"1066be82903bac4dd1e297ce1fbe3d38a9f9976b8f6d49d0a6a8e4046a0c9bc5"} Dec 06 14:15:21 crc kubenswrapper[4706]: I1206 14:15:21.751878 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"5f976297-d9ed-4fcf-a715-22d963cacb04","Type":"ContainerStarted","Data":"7b518441b88d20281e3f36bf5978aec4d98857d78a401eabc8d654b46f59f5ef"} Dec 06 14:15:21 crc kubenswrapper[4706]: I1206 14:15:21.754603 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"34735cc4-f322-45b7-b2cd-7dd16109952f","Type":"ContainerStarted","Data":"9f5a6ec31e09f209408528b2602b825cf337de499f3b0397ccc86cce2e7e870c"} Dec 06 14:15:21 crc kubenswrapper[4706]: I1206 14:15:21.754626 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"34735cc4-f322-45b7-b2cd-7dd16109952f","Type":"ContainerStarted","Data":"92f5f6a5c272d2e4efca06f4de223fc48cdde09a19888a0c8b9979696c026c4b"} Dec 06 14:15:22 crc kubenswrapper[4706]: I1206 14:15:22.769198 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"34735cc4-f322-45b7-b2cd-7dd16109952f","Type":"ContainerStarted","Data":"8f8e1e2b803d43923561ce6cced2295a09a89278e837972f7e1fc0515121eb5e"} Dec 06 14:15:22 crc kubenswrapper[4706]: I1206 14:15:22.771475 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:22 crc kubenswrapper[4706]: I1206 14:15:22.775595 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"572aefd5-0c7c-4609-9852-582b7b3071c6","Type":"ContainerStarted","Data":"fa6f4360fdeac8b66dc73584fe20009ecccd7b69a2aa3166cf4820500a5d99d8"} Dec 06 14:15:22 crc kubenswrapper[4706]: I1206 14:15:22.775900 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:22 crc kubenswrapper[4706]: I1206 14:15:22.782471 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"5f976297-d9ed-4fcf-a715-22d963cacb04","Type":"ContainerStarted","Data":"0ea6cf73df063dab5b34c8f12299cf85776ccc2539972611b9653fbfe68aea68"} Dec 06 14:15:22 crc kubenswrapper[4706]: I1206 14:15:22.782995 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:22 crc kubenswrapper[4706]: I1206 14:15:22.808565 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-0" podStartSLOduration=2.808536572 podStartE2EDuration="2.808536572s" podCreationTimestamp="2025-12-06 14:15:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:15:22.793210815 +0000 UTC m=+1165.629003722" watchObservedRunningTime="2025-12-06 14:15:22.808536572 +0000 UTC m=+1165.644329469" Dec 06 14:15:22 crc kubenswrapper[4706]: I1206 14:15:22.824607 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-1" podStartSLOduration=2.824581087 podStartE2EDuration="2.824581087s" podCreationTimestamp="2025-12-06 14:15:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:15:22.817347505 +0000 UTC m=+1165.653140412" watchObservedRunningTime="2025-12-06 14:15:22.824581087 +0000 UTC m=+1165.660373984" Dec 06 14:15:22 crc kubenswrapper[4706]: I1206 14:15:22.848875 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-2" podStartSLOduration=2.84885544 podStartE2EDuration="2.84885544s" podCreationTimestamp="2025-12-06 14:15:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:15:22.841907806 +0000 UTC m=+1165.677700673" watchObservedRunningTime="2025-12-06 14:15:22.84885544 +0000 UTC m=+1165.684648307" Dec 06 14:15:25 crc kubenswrapper[4706]: I1206 14:15:25.511444 4706 scope.go:117] "RemoveContainer" containerID="7ef4bbc8f42569aad0b46e84bb4ddecb1531e14ecef5b4eec1ad04eeb93a0b6b" Dec 06 14:15:25 crc kubenswrapper[4706]: I1206 14:15:25.512216 4706 scope.go:117] "RemoveContainer" containerID="604f2eb8b124b8739f9dada14084488fa9e69b991c9482a2fd328027bb04dbc6" Dec 06 14:15:25 crc kubenswrapper[4706]: E1206 14:15:25.512730 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" Dec 06 14:15:32 crc kubenswrapper[4706]: I1206 14:15:32.207433 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:32 crc kubenswrapper[4706]: I1206 14:15:32.329430 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:32 crc kubenswrapper[4706]: I1206 14:15:32.442763 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:33 crc kubenswrapper[4706]: I1206 14:15:33.424633 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 06 14:15:33 crc kubenswrapper[4706]: I1206 14:15:33.424869 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-2" podUID="5f976297-d9ed-4fcf-a715-22d963cacb04" containerName="cinder-api-log" containerID="cri-o://1066be82903bac4dd1e297ce1fbe3d38a9f9976b8f6d49d0a6a8e4046a0c9bc5" gracePeriod=30 Dec 06 14:15:33 crc kubenswrapper[4706]: I1206 14:15:33.425293 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-2" podUID="5f976297-d9ed-4fcf-a715-22d963cacb04" containerName="cinder-api" containerID="cri-o://0ea6cf73df063dab5b34c8f12299cf85776ccc2539972611b9653fbfe68aea68" gracePeriod=30 Dec 06 14:15:33 crc kubenswrapper[4706]: I1206 14:15:33.440899 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 06 14:15:33 crc kubenswrapper[4706]: I1206 14:15:33.441129 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-1" podUID="572aefd5-0c7c-4609-9852-582b7b3071c6" containerName="cinder-api-log" containerID="cri-o://5c4233562a3c60346725dac6f7ca868a58d2aca379fdcae21a5964b96c5b7571" gracePeriod=30 Dec 06 14:15:33 crc kubenswrapper[4706]: I1206 14:15:33.441687 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-1" podUID="572aefd5-0c7c-4609-9852-582b7b3071c6" containerName="cinder-api" containerID="cri-o://fa6f4360fdeac8b66dc73584fe20009ecccd7b69a2aa3166cf4820500a5d99d8" gracePeriod=30 Dec 06 14:15:33 crc kubenswrapper[4706]: I1206 14:15:33.441969 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="cinder-kuttl-tests/cinder-api-2" podUID="5f976297-d9ed-4fcf-a715-22d963cacb04" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.99:8776/healthcheck\": EOF" Dec 06 14:15:33 crc kubenswrapper[4706]: I1206 14:15:33.459481 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="cinder-kuttl-tests/cinder-api-1" podUID="572aefd5-0c7c-4609-9852-582b7b3071c6" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.100:8776/healthcheck\": EOF" Dec 06 14:15:33 crc kubenswrapper[4706]: I1206 14:15:33.876602 4706 generic.go:334] "Generic (PLEG): container finished" podID="572aefd5-0c7c-4609-9852-582b7b3071c6" containerID="5c4233562a3c60346725dac6f7ca868a58d2aca379fdcae21a5964b96c5b7571" exitCode=143 Dec 06 14:15:33 crc kubenswrapper[4706]: I1206 14:15:33.876654 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"572aefd5-0c7c-4609-9852-582b7b3071c6","Type":"ContainerDied","Data":"5c4233562a3c60346725dac6f7ca868a58d2aca379fdcae21a5964b96c5b7571"} Dec 06 14:15:33 crc kubenswrapper[4706]: I1206 14:15:33.879951 4706 generic.go:334] "Generic (PLEG): container finished" podID="5f976297-d9ed-4fcf-a715-22d963cacb04" containerID="1066be82903bac4dd1e297ce1fbe3d38a9f9976b8f6d49d0a6a8e4046a0c9bc5" exitCode=143 Dec 06 14:15:33 crc kubenswrapper[4706]: I1206 14:15:33.880004 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"5f976297-d9ed-4fcf-a715-22d963cacb04","Type":"ContainerDied","Data":"1066be82903bac4dd1e297ce1fbe3d38a9f9976b8f6d49d0a6a8e4046a0c9bc5"} Dec 06 14:15:36 crc kubenswrapper[4706]: I1206 14:15:36.512079 4706 scope.go:117] "RemoveContainer" containerID="7ef4bbc8f42569aad0b46e84bb4ddecb1531e14ecef5b4eec1ad04eeb93a0b6b" Dec 06 14:15:36 crc kubenswrapper[4706]: I1206 14:15:36.513409 4706 scope.go:117] "RemoveContainer" containerID="604f2eb8b124b8739f9dada14084488fa9e69b991c9482a2fd328027bb04dbc6" Dec 06 14:15:36 crc kubenswrapper[4706]: E1206 14:15:36.513702 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(49814569-1bcb-4ddd-97db-7115d5ec48d1)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" Dec 06 14:15:37 crc kubenswrapper[4706]: I1206 14:15:37.860695 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-2" podUID="5f976297-d9ed-4fcf-a715-22d963cacb04" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.99:8776/healthcheck\": read tcp 10.217.0.2:33406->10.217.0.99:8776: read: connection reset by peer" Dec 06 14:15:37 crc kubenswrapper[4706]: I1206 14:15:37.882881 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-1" podUID="572aefd5-0c7c-4609-9852-582b7b3071c6" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.100:8776/healthcheck\": read tcp 10.217.0.2:49624->10.217.0.100:8776: read: connection reset by peer" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.268048 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.272494 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.363402 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/572aefd5-0c7c-4609-9852-582b7b3071c6-logs\") pod \"572aefd5-0c7c-4609-9852-582b7b3071c6\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.363458 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f976297-d9ed-4fcf-a715-22d963cacb04-logs\") pod \"5f976297-d9ed-4fcf-a715-22d963cacb04\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.363511 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-config-data-custom\") pod \"5f976297-d9ed-4fcf-a715-22d963cacb04\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.363528 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mddgp\" (UniqueName: \"kubernetes.io/projected/572aefd5-0c7c-4609-9852-582b7b3071c6-kube-api-access-mddgp\") pod \"572aefd5-0c7c-4609-9852-582b7b3071c6\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.363551 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dn94x\" (UniqueName: \"kubernetes.io/projected/5f976297-d9ed-4fcf-a715-22d963cacb04-kube-api-access-dn94x\") pod \"5f976297-d9ed-4fcf-a715-22d963cacb04\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.363574 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-scripts\") pod \"572aefd5-0c7c-4609-9852-582b7b3071c6\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.363588 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5f976297-d9ed-4fcf-a715-22d963cacb04-etc-machine-id\") pod \"5f976297-d9ed-4fcf-a715-22d963cacb04\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.363649 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-config-data\") pod \"572aefd5-0c7c-4609-9852-582b7b3071c6\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.363672 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-config-data-custom\") pod \"572aefd5-0c7c-4609-9852-582b7b3071c6\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.363699 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-scripts\") pod \"5f976297-d9ed-4fcf-a715-22d963cacb04\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.363711 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/572aefd5-0c7c-4609-9852-582b7b3071c6-etc-machine-id\") pod \"572aefd5-0c7c-4609-9852-582b7b3071c6\" (UID: \"572aefd5-0c7c-4609-9852-582b7b3071c6\") " Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.363731 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-config-data\") pod \"5f976297-d9ed-4fcf-a715-22d963cacb04\" (UID: \"5f976297-d9ed-4fcf-a715-22d963cacb04\") " Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.364525 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f976297-d9ed-4fcf-a715-22d963cacb04-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5f976297-d9ed-4fcf-a715-22d963cacb04" (UID: "5f976297-d9ed-4fcf-a715-22d963cacb04"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.364647 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/572aefd5-0c7c-4609-9852-582b7b3071c6-logs" (OuterVolumeSpecName: "logs") pod "572aefd5-0c7c-4609-9852-582b7b3071c6" (UID: "572aefd5-0c7c-4609-9852-582b7b3071c6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.364712 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/572aefd5-0c7c-4609-9852-582b7b3071c6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "572aefd5-0c7c-4609-9852-582b7b3071c6" (UID: "572aefd5-0c7c-4609-9852-582b7b3071c6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.365844 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f976297-d9ed-4fcf-a715-22d963cacb04-logs" (OuterVolumeSpecName: "logs") pod "5f976297-d9ed-4fcf-a715-22d963cacb04" (UID: "5f976297-d9ed-4fcf-a715-22d963cacb04"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.369639 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "572aefd5-0c7c-4609-9852-582b7b3071c6" (UID: "572aefd5-0c7c-4609-9852-582b7b3071c6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.369793 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f976297-d9ed-4fcf-a715-22d963cacb04-kube-api-access-dn94x" (OuterVolumeSpecName: "kube-api-access-dn94x") pod "5f976297-d9ed-4fcf-a715-22d963cacb04" (UID: "5f976297-d9ed-4fcf-a715-22d963cacb04"). InnerVolumeSpecName "kube-api-access-dn94x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.369803 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-scripts" (OuterVolumeSpecName: "scripts") pod "572aefd5-0c7c-4609-9852-582b7b3071c6" (UID: "572aefd5-0c7c-4609-9852-582b7b3071c6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.370068 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/572aefd5-0c7c-4609-9852-582b7b3071c6-kube-api-access-mddgp" (OuterVolumeSpecName: "kube-api-access-mddgp") pod "572aefd5-0c7c-4609-9852-582b7b3071c6" (UID: "572aefd5-0c7c-4609-9852-582b7b3071c6"). InnerVolumeSpecName "kube-api-access-mddgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.370163 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-scripts" (OuterVolumeSpecName: "scripts") pod "5f976297-d9ed-4fcf-a715-22d963cacb04" (UID: "5f976297-d9ed-4fcf-a715-22d963cacb04"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.370279 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5f976297-d9ed-4fcf-a715-22d963cacb04" (UID: "5f976297-d9ed-4fcf-a715-22d963cacb04"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.396407 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-config-data" (OuterVolumeSpecName: "config-data") pod "572aefd5-0c7c-4609-9852-582b7b3071c6" (UID: "572aefd5-0c7c-4609-9852-582b7b3071c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.397020 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-config-data" (OuterVolumeSpecName: "config-data") pod "5f976297-d9ed-4fcf-a715-22d963cacb04" (UID: "5f976297-d9ed-4fcf-a715-22d963cacb04"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.464972 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.465015 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.465029 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.465042 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/572aefd5-0c7c-4609-9852-582b7b3071c6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.465054 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.465065 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/572aefd5-0c7c-4609-9852-582b7b3071c6-logs\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.465075 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f976297-d9ed-4fcf-a715-22d963cacb04-logs\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.465086 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f976297-d9ed-4fcf-a715-22d963cacb04-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.465097 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mddgp\" (UniqueName: \"kubernetes.io/projected/572aefd5-0c7c-4609-9852-582b7b3071c6-kube-api-access-mddgp\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.465109 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dn94x\" (UniqueName: \"kubernetes.io/projected/5f976297-d9ed-4fcf-a715-22d963cacb04-kube-api-access-dn94x\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.465122 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5f976297-d9ed-4fcf-a715-22d963cacb04-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.465134 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/572aefd5-0c7c-4609-9852-582b7b3071c6-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.930727 4706 generic.go:334] "Generic (PLEG): container finished" podID="5f976297-d9ed-4fcf-a715-22d963cacb04" containerID="0ea6cf73df063dab5b34c8f12299cf85776ccc2539972611b9653fbfe68aea68" exitCode=0 Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.930796 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"5f976297-d9ed-4fcf-a715-22d963cacb04","Type":"ContainerDied","Data":"0ea6cf73df063dab5b34c8f12299cf85776ccc2539972611b9653fbfe68aea68"} Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.930806 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-2" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.930984 4706 scope.go:117] "RemoveContainer" containerID="0ea6cf73df063dab5b34c8f12299cf85776ccc2539972611b9653fbfe68aea68" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.930956 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"5f976297-d9ed-4fcf-a715-22d963cacb04","Type":"ContainerDied","Data":"7b518441b88d20281e3f36bf5978aec4d98857d78a401eabc8d654b46f59f5ef"} Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.934526 4706 generic.go:334] "Generic (PLEG): container finished" podID="572aefd5-0c7c-4609-9852-582b7b3071c6" containerID="fa6f4360fdeac8b66dc73584fe20009ecccd7b69a2aa3166cf4820500a5d99d8" exitCode=0 Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.934587 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-1" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.934599 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"572aefd5-0c7c-4609-9852-582b7b3071c6","Type":"ContainerDied","Data":"fa6f4360fdeac8b66dc73584fe20009ecccd7b69a2aa3166cf4820500a5d99d8"} Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.934906 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"572aefd5-0c7c-4609-9852-582b7b3071c6","Type":"ContainerDied","Data":"7e6636685549b062083e36e81ee5b76ebee7fd2f5978499e9d564b2d42c89bbb"} Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.958842 4706 scope.go:117] "RemoveContainer" containerID="1066be82903bac4dd1e297ce1fbe3d38a9f9976b8f6d49d0a6a8e4046a0c9bc5" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.981942 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.983644 4706 scope.go:117] "RemoveContainer" containerID="0ea6cf73df063dab5b34c8f12299cf85776ccc2539972611b9653fbfe68aea68" Dec 06 14:15:38 crc kubenswrapper[4706]: E1206 14:15:38.984398 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ea6cf73df063dab5b34c8f12299cf85776ccc2539972611b9653fbfe68aea68\": container with ID starting with 0ea6cf73df063dab5b34c8f12299cf85776ccc2539972611b9653fbfe68aea68 not found: ID does not exist" containerID="0ea6cf73df063dab5b34c8f12299cf85776ccc2539972611b9653fbfe68aea68" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.984438 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ea6cf73df063dab5b34c8f12299cf85776ccc2539972611b9653fbfe68aea68"} err="failed to get container status \"0ea6cf73df063dab5b34c8f12299cf85776ccc2539972611b9653fbfe68aea68\": rpc error: code = NotFound desc = could not find container \"0ea6cf73df063dab5b34c8f12299cf85776ccc2539972611b9653fbfe68aea68\": container with ID starting with 0ea6cf73df063dab5b34c8f12299cf85776ccc2539972611b9653fbfe68aea68 not found: ID does not exist" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.984464 4706 scope.go:117] "RemoveContainer" containerID="1066be82903bac4dd1e297ce1fbe3d38a9f9976b8f6d49d0a6a8e4046a0c9bc5" Dec 06 14:15:38 crc kubenswrapper[4706]: E1206 14:15:38.990508 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1066be82903bac4dd1e297ce1fbe3d38a9f9976b8f6d49d0a6a8e4046a0c9bc5\": container with ID starting with 1066be82903bac4dd1e297ce1fbe3d38a9f9976b8f6d49d0a6a8e4046a0c9bc5 not found: ID does not exist" containerID="1066be82903bac4dd1e297ce1fbe3d38a9f9976b8f6d49d0a6a8e4046a0c9bc5" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.990560 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1066be82903bac4dd1e297ce1fbe3d38a9f9976b8f6d49d0a6a8e4046a0c9bc5"} err="failed to get container status \"1066be82903bac4dd1e297ce1fbe3d38a9f9976b8f6d49d0a6a8e4046a0c9bc5\": rpc error: code = NotFound desc = could not find container \"1066be82903bac4dd1e297ce1fbe3d38a9f9976b8f6d49d0a6a8e4046a0c9bc5\": container with ID starting with 1066be82903bac4dd1e297ce1fbe3d38a9f9976b8f6d49d0a6a8e4046a0c9bc5 not found: ID does not exist" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.990586 4706 scope.go:117] "RemoveContainer" containerID="fa6f4360fdeac8b66dc73584fe20009ecccd7b69a2aa3166cf4820500a5d99d8" Dec 06 14:15:38 crc kubenswrapper[4706]: I1206 14:15:38.996967 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.006694 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.013756 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.026496 4706 scope.go:117] "RemoveContainer" containerID="5c4233562a3c60346725dac6f7ca868a58d2aca379fdcae21a5964b96c5b7571" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.040146 4706 scope.go:117] "RemoveContainer" containerID="fa6f4360fdeac8b66dc73584fe20009ecccd7b69a2aa3166cf4820500a5d99d8" Dec 06 14:15:39 crc kubenswrapper[4706]: E1206 14:15:39.040520 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa6f4360fdeac8b66dc73584fe20009ecccd7b69a2aa3166cf4820500a5d99d8\": container with ID starting with fa6f4360fdeac8b66dc73584fe20009ecccd7b69a2aa3166cf4820500a5d99d8 not found: ID does not exist" containerID="fa6f4360fdeac8b66dc73584fe20009ecccd7b69a2aa3166cf4820500a5d99d8" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.040561 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa6f4360fdeac8b66dc73584fe20009ecccd7b69a2aa3166cf4820500a5d99d8"} err="failed to get container status \"fa6f4360fdeac8b66dc73584fe20009ecccd7b69a2aa3166cf4820500a5d99d8\": rpc error: code = NotFound desc = could not find container \"fa6f4360fdeac8b66dc73584fe20009ecccd7b69a2aa3166cf4820500a5d99d8\": container with ID starting with fa6f4360fdeac8b66dc73584fe20009ecccd7b69a2aa3166cf4820500a5d99d8 not found: ID does not exist" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.040588 4706 scope.go:117] "RemoveContainer" containerID="5c4233562a3c60346725dac6f7ca868a58d2aca379fdcae21a5964b96c5b7571" Dec 06 14:15:39 crc kubenswrapper[4706]: E1206 14:15:39.040918 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c4233562a3c60346725dac6f7ca868a58d2aca379fdcae21a5964b96c5b7571\": container with ID starting with 5c4233562a3c60346725dac6f7ca868a58d2aca379fdcae21a5964b96c5b7571 not found: ID does not exist" containerID="5c4233562a3c60346725dac6f7ca868a58d2aca379fdcae21a5964b96c5b7571" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.040945 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c4233562a3c60346725dac6f7ca868a58d2aca379fdcae21a5964b96c5b7571"} err="failed to get container status \"5c4233562a3c60346725dac6f7ca868a58d2aca379fdcae21a5964b96c5b7571\": rpc error: code = NotFound desc = could not find container \"5c4233562a3c60346725dac6f7ca868a58d2aca379fdcae21a5964b96c5b7571\": container with ID starting with 5c4233562a3c60346725dac6f7ca868a58d2aca379fdcae21a5964b96c5b7571 not found: ID does not exist" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.529603 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="572aefd5-0c7c-4609-9852-582b7b3071c6" path="/var/lib/kubelet/pods/572aefd5-0c7c-4609-9852-582b7b3071c6/volumes" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.531034 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f976297-d9ed-4fcf-a715-22d963cacb04" path="/var/lib/kubelet/pods/5f976297-d9ed-4fcf-a715-22d963cacb04/volumes" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.723169 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-ndc2g"] Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.731160 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-ndc2g"] Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.751623 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.760001 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.782395 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.782688 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="b77d62e6-8988-4135-a645-14fcdd3d9c92" containerName="cinder-backup" containerID="cri-o://98c8c2349e48854c0ef7f4fd6f2a3ac4dbc581f868848b433127270a7dca0cab" gracePeriod=30 Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.782762 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="b77d62e6-8988-4135-a645-14fcdd3d9c92" containerName="probe" containerID="cri-o://6734078a417bc2c57cba55d9833e2b2d9915ce3ce5c679df08017c80123fa482" gracePeriod=30 Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.793261 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder06a5-account-delete-wrhnv"] Dec 06 14:15:39 crc kubenswrapper[4706]: E1206 14:15:39.793617 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f976297-d9ed-4fcf-a715-22d963cacb04" containerName="cinder-api" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.793634 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f976297-d9ed-4fcf-a715-22d963cacb04" containerName="cinder-api" Dec 06 14:15:39 crc kubenswrapper[4706]: E1206 14:15:39.793645 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572aefd5-0c7c-4609-9852-582b7b3071c6" containerName="cinder-api" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.793653 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="572aefd5-0c7c-4609-9852-582b7b3071c6" containerName="cinder-api" Dec 06 14:15:39 crc kubenswrapper[4706]: E1206 14:15:39.793665 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f976297-d9ed-4fcf-a715-22d963cacb04" containerName="cinder-api-log" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.793681 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f976297-d9ed-4fcf-a715-22d963cacb04" containerName="cinder-api-log" Dec 06 14:15:39 crc kubenswrapper[4706]: E1206 14:15:39.793691 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572aefd5-0c7c-4609-9852-582b7b3071c6" containerName="cinder-api-log" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.793698 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="572aefd5-0c7c-4609-9852-582b7b3071c6" containerName="cinder-api-log" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.793906 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f976297-d9ed-4fcf-a715-22d963cacb04" containerName="cinder-api" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.793918 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="572aefd5-0c7c-4609-9852-582b7b3071c6" containerName="cinder-api-log" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.793927 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="572aefd5-0c7c-4609-9852-582b7b3071c6" containerName="cinder-api" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.793946 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f976297-d9ed-4fcf-a715-22d963cacb04" containerName="cinder-api-log" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.794492 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder06a5-account-delete-wrhnv" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.804234 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.804543 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="34735cc4-f322-45b7-b2cd-7dd16109952f" containerName="cinder-api-log" containerID="cri-o://9f5a6ec31e09f209408528b2602b825cf337de499f3b0397ccc86cce2e7e870c" gracePeriod=30 Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.804704 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="34735cc4-f322-45b7-b2cd-7dd16109952f" containerName="cinder-api" containerID="cri-o://8f8e1e2b803d43923561ce6cced2295a09a89278e837972f7e1fc0515121eb5e" gracePeriod=30 Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.814747 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder06a5-account-delete-wrhnv"] Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.886371 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfsjq\" (UniqueName: \"kubernetes.io/projected/d3335ed5-4a44-4d31-ac61-73abbf2a8c18-kube-api-access-mfsjq\") pod \"cinder06a5-account-delete-wrhnv\" (UID: \"d3335ed5-4a44-4d31-ac61-73abbf2a8c18\") " pod="cinder-kuttl-tests/cinder06a5-account-delete-wrhnv" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.886503 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3335ed5-4a44-4d31-ac61-73abbf2a8c18-operator-scripts\") pod \"cinder06a5-account-delete-wrhnv\" (UID: \"d3335ed5-4a44-4d31-ac61-73abbf2a8c18\") " pod="cinder-kuttl-tests/cinder06a5-account-delete-wrhnv" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.948427 4706 generic.go:334] "Generic (PLEG): container finished" podID="34735cc4-f322-45b7-b2cd-7dd16109952f" containerID="9f5a6ec31e09f209408528b2602b825cf337de499f3b0397ccc86cce2e7e870c" exitCode=143 Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.948561 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"34735cc4-f322-45b7-b2cd-7dd16109952f","Type":"ContainerDied","Data":"9f5a6ec31e09f209408528b2602b825cf337de499f3b0397ccc86cce2e7e870c"} Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.951834 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="2316de26-2427-42bf-9d27-f84217c2c343" containerName="cinder-scheduler" containerID="cri-o://1b6ce9a5b2d08ad2637a9d0ee2c9cb361dbb07b4118e62737cfd8efb87ccce59" gracePeriod=30 Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.951942 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="2316de26-2427-42bf-9d27-f84217c2c343" containerName="probe" containerID="cri-o://32f0fc2b98f942acc226f0168315800abf4fa319ea4460af717ccd743092710c" gracePeriod=30 Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.988057 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfsjq\" (UniqueName: \"kubernetes.io/projected/d3335ed5-4a44-4d31-ac61-73abbf2a8c18-kube-api-access-mfsjq\") pod \"cinder06a5-account-delete-wrhnv\" (UID: \"d3335ed5-4a44-4d31-ac61-73abbf2a8c18\") " pod="cinder-kuttl-tests/cinder06a5-account-delete-wrhnv" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.988196 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3335ed5-4a44-4d31-ac61-73abbf2a8c18-operator-scripts\") pod \"cinder06a5-account-delete-wrhnv\" (UID: \"d3335ed5-4a44-4d31-ac61-73abbf2a8c18\") " pod="cinder-kuttl-tests/cinder06a5-account-delete-wrhnv" Dec 06 14:15:39 crc kubenswrapper[4706]: I1206 14:15:39.989000 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3335ed5-4a44-4d31-ac61-73abbf2a8c18-operator-scripts\") pod \"cinder06a5-account-delete-wrhnv\" (UID: \"d3335ed5-4a44-4d31-ac61-73abbf2a8c18\") " pod="cinder-kuttl-tests/cinder06a5-account-delete-wrhnv" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.006928 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfsjq\" (UniqueName: \"kubernetes.io/projected/d3335ed5-4a44-4d31-ac61-73abbf2a8c18-kube-api-access-mfsjq\") pod \"cinder06a5-account-delete-wrhnv\" (UID: \"d3335ed5-4a44-4d31-ac61-73abbf2a8c18\") " pod="cinder-kuttl-tests/cinder06a5-account-delete-wrhnv" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.056238 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.125882 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder06a5-account-delete-wrhnv" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.190237 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-run\") pod \"49814569-1bcb-4ddd-97db-7115d5ec48d1\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.190416 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-locks-cinder\") pod \"49814569-1bcb-4ddd-97db-7115d5ec48d1\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.190362 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-run" (OuterVolumeSpecName: "run") pod "49814569-1bcb-4ddd-97db-7115d5ec48d1" (UID: "49814569-1bcb-4ddd-97db-7115d5ec48d1"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.190513 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vp2w5\" (UniqueName: \"kubernetes.io/projected/49814569-1bcb-4ddd-97db-7115d5ec48d1-kube-api-access-vp2w5\") pod \"49814569-1bcb-4ddd-97db-7115d5ec48d1\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.190562 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-dev\") pod \"49814569-1bcb-4ddd-97db-7115d5ec48d1\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.190582 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-iscsi\") pod \"49814569-1bcb-4ddd-97db-7115d5ec48d1\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.191165 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-lib-cinder\") pod \"49814569-1bcb-4ddd-97db-7115d5ec48d1\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.191215 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-sys\") pod \"49814569-1bcb-4ddd-97db-7115d5ec48d1\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.191271 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-lib-modules\") pod \"49814569-1bcb-4ddd-97db-7115d5ec48d1\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.191371 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-config-data\") pod \"49814569-1bcb-4ddd-97db-7115d5ec48d1\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.191399 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-locks-brick\") pod \"49814569-1bcb-4ddd-97db-7115d5ec48d1\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.191440 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-config-data-custom\") pod \"49814569-1bcb-4ddd-97db-7115d5ec48d1\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.191477 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-nvme\") pod \"49814569-1bcb-4ddd-97db-7115d5ec48d1\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.191507 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-scripts\") pod \"49814569-1bcb-4ddd-97db-7115d5ec48d1\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.191532 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-machine-id\") pod \"49814569-1bcb-4ddd-97db-7115d5ec48d1\" (UID: \"49814569-1bcb-4ddd-97db-7115d5ec48d1\") " Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.192015 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-run\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.190593 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "49814569-1bcb-4ddd-97db-7115d5ec48d1" (UID: "49814569-1bcb-4ddd-97db-7115d5ec48d1"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.190624 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-dev" (OuterVolumeSpecName: "dev") pod "49814569-1bcb-4ddd-97db-7115d5ec48d1" (UID: "49814569-1bcb-4ddd-97db-7115d5ec48d1"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.190721 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "49814569-1bcb-4ddd-97db-7115d5ec48d1" (UID: "49814569-1bcb-4ddd-97db-7115d5ec48d1"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.192062 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "49814569-1bcb-4ddd-97db-7115d5ec48d1" (UID: "49814569-1bcb-4ddd-97db-7115d5ec48d1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.192117 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "49814569-1bcb-4ddd-97db-7115d5ec48d1" (UID: "49814569-1bcb-4ddd-97db-7115d5ec48d1"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.192141 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-sys" (OuterVolumeSpecName: "sys") pod "49814569-1bcb-4ddd-97db-7115d5ec48d1" (UID: "49814569-1bcb-4ddd-97db-7115d5ec48d1"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.192162 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "49814569-1bcb-4ddd-97db-7115d5ec48d1" (UID: "49814569-1bcb-4ddd-97db-7115d5ec48d1"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.192801 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "49814569-1bcb-4ddd-97db-7115d5ec48d1" (UID: "49814569-1bcb-4ddd-97db-7115d5ec48d1"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.192845 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "49814569-1bcb-4ddd-97db-7115d5ec48d1" (UID: "49814569-1bcb-4ddd-97db-7115d5ec48d1"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.194463 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49814569-1bcb-4ddd-97db-7115d5ec48d1-kube-api-access-vp2w5" (OuterVolumeSpecName: "kube-api-access-vp2w5") pod "49814569-1bcb-4ddd-97db-7115d5ec48d1" (UID: "49814569-1bcb-4ddd-97db-7115d5ec48d1"). InnerVolumeSpecName "kube-api-access-vp2w5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.196586 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "49814569-1bcb-4ddd-97db-7115d5ec48d1" (UID: "49814569-1bcb-4ddd-97db-7115d5ec48d1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.197374 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-scripts" (OuterVolumeSpecName: "scripts") pod "49814569-1bcb-4ddd-97db-7115d5ec48d1" (UID: "49814569-1bcb-4ddd-97db-7115d5ec48d1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.271680 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-config-data" (OuterVolumeSpecName: "config-data") pod "49814569-1bcb-4ddd-97db-7115d5ec48d1" (UID: "49814569-1bcb-4ddd-97db-7115d5ec48d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.293290 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.293577 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vp2w5\" (UniqueName: \"kubernetes.io/projected/49814569-1bcb-4ddd-97db-7115d5ec48d1-kube-api-access-vp2w5\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.293590 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-dev\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.293600 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.293608 4706 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.293618 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-sys\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.293625 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.293633 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.293642 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.293651 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.293659 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.293667 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49814569-1bcb-4ddd-97db-7115d5ec48d1-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.293675 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/49814569-1bcb-4ddd-97db-7115d5ec48d1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.586767 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder06a5-account-delete-wrhnv"] Dec 06 14:15:40 crc kubenswrapper[4706]: W1206 14:15:40.588394 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3335ed5_4a44_4d31_ac61_73abbf2a8c18.slice/crio-f4e12fabf3d65fdd958beab31dfef7d07d1a8eb099bd487fa612796754d2da3c WatchSource:0}: Error finding container f4e12fabf3d65fdd958beab31dfef7d07d1a8eb099bd487fa612796754d2da3c: Status 404 returned error can't find the container with id f4e12fabf3d65fdd958beab31dfef7d07d1a8eb099bd487fa612796754d2da3c Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.963371 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.963407 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"49814569-1bcb-4ddd-97db-7115d5ec48d1","Type":"ContainerDied","Data":"45595cc5fe6d491c706a93f685f0990f79af0d1eb411e6c53ebbad86f2458c54"} Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.963454 4706 scope.go:117] "RemoveContainer" containerID="604f2eb8b124b8739f9dada14084488fa9e69b991c9482a2fd328027bb04dbc6" Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.965906 4706 generic.go:334] "Generic (PLEG): container finished" podID="2316de26-2427-42bf-9d27-f84217c2c343" containerID="32f0fc2b98f942acc226f0168315800abf4fa319ea4460af717ccd743092710c" exitCode=0 Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.966009 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"2316de26-2427-42bf-9d27-f84217c2c343","Type":"ContainerDied","Data":"32f0fc2b98f942acc226f0168315800abf4fa319ea4460af717ccd743092710c"} Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.968477 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder06a5-account-delete-wrhnv" event={"ID":"d3335ed5-4a44-4d31-ac61-73abbf2a8c18","Type":"ContainerStarted","Data":"13808474ccfd99d7681ce4763e54a102d84d3cc2e4965084e152229020c4fea1"} Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.968514 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder06a5-account-delete-wrhnv" event={"ID":"d3335ed5-4a44-4d31-ac61-73abbf2a8c18","Type":"ContainerStarted","Data":"f4e12fabf3d65fdd958beab31dfef7d07d1a8eb099bd487fa612796754d2da3c"} Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.973719 4706 generic.go:334] "Generic (PLEG): container finished" podID="b77d62e6-8988-4135-a645-14fcdd3d9c92" containerID="6734078a417bc2c57cba55d9833e2b2d9915ce3ce5c679df08017c80123fa482" exitCode=0 Dec 06 14:15:40 crc kubenswrapper[4706]: I1206 14:15:40.973757 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"b77d62e6-8988-4135-a645-14fcdd3d9c92","Type":"ContainerDied","Data":"6734078a417bc2c57cba55d9833e2b2d9915ce3ce5c679df08017c80123fa482"} Dec 06 14:15:41 crc kubenswrapper[4706]: I1206 14:15:41.007462 4706 scope.go:117] "RemoveContainer" containerID="7ef4bbc8f42569aad0b46e84bb4ddecb1531e14ecef5b4eec1ad04eeb93a0b6b" Dec 06 14:15:41 crc kubenswrapper[4706]: I1206 14:15:41.012184 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder06a5-account-delete-wrhnv" podStartSLOduration=2.012170135 podStartE2EDuration="2.012170135s" podCreationTimestamp="2025-12-06 14:15:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:15:40.991917618 +0000 UTC m=+1183.827710475" watchObservedRunningTime="2025-12-06 14:15:41.012170135 +0000 UTC m=+1183.847962992" Dec 06 14:15:41 crc kubenswrapper[4706]: I1206 14:15:41.023991 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 06 14:15:41 crc kubenswrapper[4706]: I1206 14:15:41.029362 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 06 14:15:41 crc kubenswrapper[4706]: I1206 14:15:41.528375 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" path="/var/lib/kubelet/pods/49814569-1bcb-4ddd-97db-7115d5ec48d1/volumes" Dec 06 14:15:41 crc kubenswrapper[4706]: I1206 14:15:41.529803 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f55d2950-f0b5-4a23-89b3-f8df608db21f" path="/var/lib/kubelet/pods/f55d2950-f0b5-4a23-89b3-f8df608db21f/volumes" Dec 06 14:15:41 crc kubenswrapper[4706]: I1206 14:15:41.986272 4706 generic.go:334] "Generic (PLEG): container finished" podID="2316de26-2427-42bf-9d27-f84217c2c343" containerID="1b6ce9a5b2d08ad2637a9d0ee2c9cb361dbb07b4118e62737cfd8efb87ccce59" exitCode=0 Dec 06 14:15:41 crc kubenswrapper[4706]: I1206 14:15:41.986345 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"2316de26-2427-42bf-9d27-f84217c2c343","Type":"ContainerDied","Data":"1b6ce9a5b2d08ad2637a9d0ee2c9cb361dbb07b4118e62737cfd8efb87ccce59"} Dec 06 14:15:41 crc kubenswrapper[4706]: I1206 14:15:41.988419 4706 generic.go:334] "Generic (PLEG): container finished" podID="d3335ed5-4a44-4d31-ac61-73abbf2a8c18" containerID="13808474ccfd99d7681ce4763e54a102d84d3cc2e4965084e152229020c4fea1" exitCode=0 Dec 06 14:15:41 crc kubenswrapper[4706]: I1206 14:15:41.988442 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder06a5-account-delete-wrhnv" event={"ID":"d3335ed5-4a44-4d31-ac61-73abbf2a8c18","Type":"ContainerDied","Data":"13808474ccfd99d7681ce4763e54a102d84d3cc2e4965084e152229020c4fea1"} Dec 06 14:15:42 crc kubenswrapper[4706]: I1206 14:15:42.293673 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:42 crc kubenswrapper[4706]: I1206 14:15:42.325928 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-config-data-custom\") pod \"2316de26-2427-42bf-9d27-f84217c2c343\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " Dec 06 14:15:42 crc kubenswrapper[4706]: I1206 14:15:42.326022 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-scripts\") pod \"2316de26-2427-42bf-9d27-f84217c2c343\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " Dec 06 14:15:42 crc kubenswrapper[4706]: I1206 14:15:42.326072 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-config-data\") pod \"2316de26-2427-42bf-9d27-f84217c2c343\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " Dec 06 14:15:42 crc kubenswrapper[4706]: I1206 14:15:42.326161 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwqnw\" (UniqueName: \"kubernetes.io/projected/2316de26-2427-42bf-9d27-f84217c2c343-kube-api-access-kwqnw\") pod \"2316de26-2427-42bf-9d27-f84217c2c343\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " Dec 06 14:15:42 crc kubenswrapper[4706]: I1206 14:15:42.326191 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2316de26-2427-42bf-9d27-f84217c2c343-etc-machine-id\") pod \"2316de26-2427-42bf-9d27-f84217c2c343\" (UID: \"2316de26-2427-42bf-9d27-f84217c2c343\") " Dec 06 14:15:42 crc kubenswrapper[4706]: I1206 14:15:42.326534 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2316de26-2427-42bf-9d27-f84217c2c343-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2316de26-2427-42bf-9d27-f84217c2c343" (UID: "2316de26-2427-42bf-9d27-f84217c2c343"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:42 crc kubenswrapper[4706]: I1206 14:15:42.338552 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-scripts" (OuterVolumeSpecName: "scripts") pod "2316de26-2427-42bf-9d27-f84217c2c343" (UID: "2316de26-2427-42bf-9d27-f84217c2c343"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:42 crc kubenswrapper[4706]: I1206 14:15:42.338576 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2316de26-2427-42bf-9d27-f84217c2c343" (UID: "2316de26-2427-42bf-9d27-f84217c2c343"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:42 crc kubenswrapper[4706]: I1206 14:15:42.338610 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2316de26-2427-42bf-9d27-f84217c2c343-kube-api-access-kwqnw" (OuterVolumeSpecName: "kube-api-access-kwqnw") pod "2316de26-2427-42bf-9d27-f84217c2c343" (UID: "2316de26-2427-42bf-9d27-f84217c2c343"). InnerVolumeSpecName "kube-api-access-kwqnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:15:42 crc kubenswrapper[4706]: I1206 14:15:42.417246 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-config-data" (OuterVolumeSpecName: "config-data") pod "2316de26-2427-42bf-9d27-f84217c2c343" (UID: "2316de26-2427-42bf-9d27-f84217c2c343"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:42 crc kubenswrapper[4706]: I1206 14:15:42.428744 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwqnw\" (UniqueName: \"kubernetes.io/projected/2316de26-2427-42bf-9d27-f84217c2c343-kube-api-access-kwqnw\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:42 crc kubenswrapper[4706]: I1206 14:15:42.428812 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2316de26-2427-42bf-9d27-f84217c2c343-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:42 crc kubenswrapper[4706]: I1206 14:15:42.428841 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:42 crc kubenswrapper[4706]: I1206 14:15:42.428866 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:42 crc kubenswrapper[4706]: I1206 14:15:42.428890 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2316de26-2427-42bf-9d27-f84217c2c343-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:42 crc kubenswrapper[4706]: I1206 14:15:42.961778 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-0" podUID="34735cc4-f322-45b7-b2cd-7dd16109952f" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.98:8776/healthcheck\": read tcp 10.217.0.2:58328->10.217.0.98:8776: read: connection reset by peer" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.000270 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"2316de26-2427-42bf-9d27-f84217c2c343","Type":"ContainerDied","Data":"f2e6eb9099f4a4866e6b6553f8dcb509112971aefa01253ff5cc844b0f5607e4"} Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.000344 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.000397 4706 scope.go:117] "RemoveContainer" containerID="32f0fc2b98f942acc226f0168315800abf4fa319ea4460af717ccd743092710c" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.064191 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.072416 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.075768 4706 scope.go:117] "RemoveContainer" containerID="1b6ce9a5b2d08ad2637a9d0ee2c9cb361dbb07b4118e62737cfd8efb87ccce59" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.288198 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder06a5-account-delete-wrhnv" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.351396 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfsjq\" (UniqueName: \"kubernetes.io/projected/d3335ed5-4a44-4d31-ac61-73abbf2a8c18-kube-api-access-mfsjq\") pod \"d3335ed5-4a44-4d31-ac61-73abbf2a8c18\" (UID: \"d3335ed5-4a44-4d31-ac61-73abbf2a8c18\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.351491 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3335ed5-4a44-4d31-ac61-73abbf2a8c18-operator-scripts\") pod \"d3335ed5-4a44-4d31-ac61-73abbf2a8c18\" (UID: \"d3335ed5-4a44-4d31-ac61-73abbf2a8c18\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.352856 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3335ed5-4a44-4d31-ac61-73abbf2a8c18-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3335ed5-4a44-4d31-ac61-73abbf2a8c18" (UID: "d3335ed5-4a44-4d31-ac61-73abbf2a8c18"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.356943 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3335ed5-4a44-4d31-ac61-73abbf2a8c18-kube-api-access-mfsjq" (OuterVolumeSpecName: "kube-api-access-mfsjq") pod "d3335ed5-4a44-4d31-ac61-73abbf2a8c18" (UID: "d3335ed5-4a44-4d31-ac61-73abbf2a8c18"). InnerVolumeSpecName "kube-api-access-mfsjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.412997 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.452866 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34735cc4-f322-45b7-b2cd-7dd16109952f-etc-machine-id\") pod \"34735cc4-f322-45b7-b2cd-7dd16109952f\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.452928 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-config-data-custom\") pod \"34735cc4-f322-45b7-b2cd-7dd16109952f\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.453032 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-scripts\") pod \"34735cc4-f322-45b7-b2cd-7dd16109952f\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.453065 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5hhr\" (UniqueName: \"kubernetes.io/projected/34735cc4-f322-45b7-b2cd-7dd16109952f-kube-api-access-d5hhr\") pod \"34735cc4-f322-45b7-b2cd-7dd16109952f\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.453091 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-config-data\") pod \"34735cc4-f322-45b7-b2cd-7dd16109952f\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.453153 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34735cc4-f322-45b7-b2cd-7dd16109952f-logs\") pod \"34735cc4-f322-45b7-b2cd-7dd16109952f\" (UID: \"34735cc4-f322-45b7-b2cd-7dd16109952f\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.453449 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfsjq\" (UniqueName: \"kubernetes.io/projected/d3335ed5-4a44-4d31-ac61-73abbf2a8c18-kube-api-access-mfsjq\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.453466 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3335ed5-4a44-4d31-ac61-73abbf2a8c18-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.453827 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34735cc4-f322-45b7-b2cd-7dd16109952f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "34735cc4-f322-45b7-b2cd-7dd16109952f" (UID: "34735cc4-f322-45b7-b2cd-7dd16109952f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.454123 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34735cc4-f322-45b7-b2cd-7dd16109952f-logs" (OuterVolumeSpecName: "logs") pod "34735cc4-f322-45b7-b2cd-7dd16109952f" (UID: "34735cc4-f322-45b7-b2cd-7dd16109952f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.456238 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-scripts" (OuterVolumeSpecName: "scripts") pod "34735cc4-f322-45b7-b2cd-7dd16109952f" (UID: "34735cc4-f322-45b7-b2cd-7dd16109952f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.456334 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "34735cc4-f322-45b7-b2cd-7dd16109952f" (UID: "34735cc4-f322-45b7-b2cd-7dd16109952f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.456399 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34735cc4-f322-45b7-b2cd-7dd16109952f-kube-api-access-d5hhr" (OuterVolumeSpecName: "kube-api-access-d5hhr") pod "34735cc4-f322-45b7-b2cd-7dd16109952f" (UID: "34735cc4-f322-45b7-b2cd-7dd16109952f"). InnerVolumeSpecName "kube-api-access-d5hhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.487845 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-config-data" (OuterVolumeSpecName: "config-data") pod "34735cc4-f322-45b7-b2cd-7dd16109952f" (UID: "34735cc4-f322-45b7-b2cd-7dd16109952f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.520515 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2316de26-2427-42bf-9d27-f84217c2c343" path="/var/lib/kubelet/pods/2316de26-2427-42bf-9d27-f84217c2c343/volumes" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.555883 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.555925 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5hhr\" (UniqueName: \"kubernetes.io/projected/34735cc4-f322-45b7-b2cd-7dd16109952f-kube-api-access-d5hhr\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.555945 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.555960 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34735cc4-f322-45b7-b2cd-7dd16109952f-logs\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.555974 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34735cc4-f322-45b7-b2cd-7dd16109952f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.555986 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34735cc4-f322-45b7-b2cd-7dd16109952f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.856987 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962145 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-config-data\") pod \"b77d62e6-8988-4135-a645-14fcdd3d9c92\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962197 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-lib-cinder\") pod \"b77d62e6-8988-4135-a645-14fcdd3d9c92\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962238 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlzsj\" (UniqueName: \"kubernetes.io/projected/b77d62e6-8988-4135-a645-14fcdd3d9c92-kube-api-access-hlzsj\") pod \"b77d62e6-8988-4135-a645-14fcdd3d9c92\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962260 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-iscsi\") pod \"b77d62e6-8988-4135-a645-14fcdd3d9c92\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962294 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-config-data-custom\") pod \"b77d62e6-8988-4135-a645-14fcdd3d9c92\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962287 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "b77d62e6-8988-4135-a645-14fcdd3d9c92" (UID: "b77d62e6-8988-4135-a645-14fcdd3d9c92"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962388 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "b77d62e6-8988-4135-a645-14fcdd3d9c92" (UID: "b77d62e6-8988-4135-a645-14fcdd3d9c92"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962416 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-nvme\") pod \"b77d62e6-8988-4135-a645-14fcdd3d9c92\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962466 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "b77d62e6-8988-4135-a645-14fcdd3d9c92" (UID: "b77d62e6-8988-4135-a645-14fcdd3d9c92"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962532 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-locks-brick\") pod \"b77d62e6-8988-4135-a645-14fcdd3d9c92\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962581 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-scripts\") pod \"b77d62e6-8988-4135-a645-14fcdd3d9c92\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962613 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-sys\") pod \"b77d62e6-8988-4135-a645-14fcdd3d9c92\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962639 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-machine-id\") pod \"b77d62e6-8988-4135-a645-14fcdd3d9c92\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962645 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "b77d62e6-8988-4135-a645-14fcdd3d9c92" (UID: "b77d62e6-8988-4135-a645-14fcdd3d9c92"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962664 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-lib-modules\") pod \"b77d62e6-8988-4135-a645-14fcdd3d9c92\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962678 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-sys" (OuterVolumeSpecName: "sys") pod "b77d62e6-8988-4135-a645-14fcdd3d9c92" (UID: "b77d62e6-8988-4135-a645-14fcdd3d9c92"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962702 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b77d62e6-8988-4135-a645-14fcdd3d9c92" (UID: "b77d62e6-8988-4135-a645-14fcdd3d9c92"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962712 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-locks-cinder\") pod \"b77d62e6-8988-4135-a645-14fcdd3d9c92\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962732 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "b77d62e6-8988-4135-a645-14fcdd3d9c92" (UID: "b77d62e6-8988-4135-a645-14fcdd3d9c92"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962751 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-dev\") pod \"b77d62e6-8988-4135-a645-14fcdd3d9c92\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962771 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-dev" (OuterVolumeSpecName: "dev") pod "b77d62e6-8988-4135-a645-14fcdd3d9c92" (UID: "b77d62e6-8988-4135-a645-14fcdd3d9c92"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962784 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-run\") pod \"b77d62e6-8988-4135-a645-14fcdd3d9c92\" (UID: \"b77d62e6-8988-4135-a645-14fcdd3d9c92\") " Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962869 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-run" (OuterVolumeSpecName: "run") pod "b77d62e6-8988-4135-a645-14fcdd3d9c92" (UID: "b77d62e6-8988-4135-a645-14fcdd3d9c92"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.962838 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "b77d62e6-8988-4135-a645-14fcdd3d9c92" (UID: "b77d62e6-8988-4135-a645-14fcdd3d9c92"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.963196 4706 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.963213 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.963223 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.963233 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.963242 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-sys\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.963252 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.963260 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.963269 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.963278 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-dev\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.963286 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b77d62e6-8988-4135-a645-14fcdd3d9c92-run\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.966579 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-scripts" (OuterVolumeSpecName: "scripts") pod "b77d62e6-8988-4135-a645-14fcdd3d9c92" (UID: "b77d62e6-8988-4135-a645-14fcdd3d9c92"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.968012 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b77d62e6-8988-4135-a645-14fcdd3d9c92-kube-api-access-hlzsj" (OuterVolumeSpecName: "kube-api-access-hlzsj") pod "b77d62e6-8988-4135-a645-14fcdd3d9c92" (UID: "b77d62e6-8988-4135-a645-14fcdd3d9c92"). InnerVolumeSpecName "kube-api-access-hlzsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:15:43 crc kubenswrapper[4706]: I1206 14:15:43.968147 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b77d62e6-8988-4135-a645-14fcdd3d9c92" (UID: "b77d62e6-8988-4135-a645-14fcdd3d9c92"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.009755 4706 generic.go:334] "Generic (PLEG): container finished" podID="34735cc4-f322-45b7-b2cd-7dd16109952f" containerID="8f8e1e2b803d43923561ce6cced2295a09a89278e837972f7e1fc0515121eb5e" exitCode=0 Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.009791 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"34735cc4-f322-45b7-b2cd-7dd16109952f","Type":"ContainerDied","Data":"8f8e1e2b803d43923561ce6cced2295a09a89278e837972f7e1fc0515121eb5e"} Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.009841 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"34735cc4-f322-45b7-b2cd-7dd16109952f","Type":"ContainerDied","Data":"92f5f6a5c272d2e4efca06f4de223fc48cdde09a19888a0c8b9979696c026c4b"} Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.009870 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.009879 4706 scope.go:117] "RemoveContainer" containerID="8f8e1e2b803d43923561ce6cced2295a09a89278e837972f7e1fc0515121eb5e" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.013480 4706 generic.go:334] "Generic (PLEG): container finished" podID="b77d62e6-8988-4135-a645-14fcdd3d9c92" containerID="98c8c2349e48854c0ef7f4fd6f2a3ac4dbc581f868848b433127270a7dca0cab" exitCode=0 Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.013548 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"b77d62e6-8988-4135-a645-14fcdd3d9c92","Type":"ContainerDied","Data":"98c8c2349e48854c0ef7f4fd6f2a3ac4dbc581f868848b433127270a7dca0cab"} Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.013586 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.013636 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"b77d62e6-8988-4135-a645-14fcdd3d9c92","Type":"ContainerDied","Data":"fcc0ead26c4fab3c5763716207701c59a263e0602d193a6d9957e84935c58b2d"} Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.017426 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder06a5-account-delete-wrhnv" event={"ID":"d3335ed5-4a44-4d31-ac61-73abbf2a8c18","Type":"ContainerDied","Data":"f4e12fabf3d65fdd958beab31dfef7d07d1a8eb099bd487fa612796754d2da3c"} Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.017630 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4e12fabf3d65fdd958beab31dfef7d07d1a8eb099bd487fa612796754d2da3c" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.017746 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder06a5-account-delete-wrhnv" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.034095 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-config-data" (OuterVolumeSpecName: "config-data") pod "b77d62e6-8988-4135-a645-14fcdd3d9c92" (UID: "b77d62e6-8988-4135-a645-14fcdd3d9c92"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.037808 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.043244 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.045292 4706 scope.go:117] "RemoveContainer" containerID="9f5a6ec31e09f209408528b2602b825cf337de499f3b0397ccc86cce2e7e870c" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.064261 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.064311 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.064353 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlzsj\" (UniqueName: \"kubernetes.io/projected/b77d62e6-8988-4135-a645-14fcdd3d9c92-kube-api-access-hlzsj\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.064369 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b77d62e6-8988-4135-a645-14fcdd3d9c92-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.065565 4706 scope.go:117] "RemoveContainer" containerID="8f8e1e2b803d43923561ce6cced2295a09a89278e837972f7e1fc0515121eb5e" Dec 06 14:15:44 crc kubenswrapper[4706]: E1206 14:15:44.066323 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f8e1e2b803d43923561ce6cced2295a09a89278e837972f7e1fc0515121eb5e\": container with ID starting with 8f8e1e2b803d43923561ce6cced2295a09a89278e837972f7e1fc0515121eb5e not found: ID does not exist" containerID="8f8e1e2b803d43923561ce6cced2295a09a89278e837972f7e1fc0515121eb5e" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.066380 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f8e1e2b803d43923561ce6cced2295a09a89278e837972f7e1fc0515121eb5e"} err="failed to get container status \"8f8e1e2b803d43923561ce6cced2295a09a89278e837972f7e1fc0515121eb5e\": rpc error: code = NotFound desc = could not find container \"8f8e1e2b803d43923561ce6cced2295a09a89278e837972f7e1fc0515121eb5e\": container with ID starting with 8f8e1e2b803d43923561ce6cced2295a09a89278e837972f7e1fc0515121eb5e not found: ID does not exist" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.066411 4706 scope.go:117] "RemoveContainer" containerID="9f5a6ec31e09f209408528b2602b825cf337de499f3b0397ccc86cce2e7e870c" Dec 06 14:15:44 crc kubenswrapper[4706]: E1206 14:15:44.066843 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f5a6ec31e09f209408528b2602b825cf337de499f3b0397ccc86cce2e7e870c\": container with ID starting with 9f5a6ec31e09f209408528b2602b825cf337de499f3b0397ccc86cce2e7e870c not found: ID does not exist" containerID="9f5a6ec31e09f209408528b2602b825cf337de499f3b0397ccc86cce2e7e870c" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.066876 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f5a6ec31e09f209408528b2602b825cf337de499f3b0397ccc86cce2e7e870c"} err="failed to get container status \"9f5a6ec31e09f209408528b2602b825cf337de499f3b0397ccc86cce2e7e870c\": rpc error: code = NotFound desc = could not find container \"9f5a6ec31e09f209408528b2602b825cf337de499f3b0397ccc86cce2e7e870c\": container with ID starting with 9f5a6ec31e09f209408528b2602b825cf337de499f3b0397ccc86cce2e7e870c not found: ID does not exist" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.066904 4706 scope.go:117] "RemoveContainer" containerID="6734078a417bc2c57cba55d9833e2b2d9915ce3ce5c679df08017c80123fa482" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.089994 4706 scope.go:117] "RemoveContainer" containerID="98c8c2349e48854c0ef7f4fd6f2a3ac4dbc581f868848b433127270a7dca0cab" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.109882 4706 scope.go:117] "RemoveContainer" containerID="6734078a417bc2c57cba55d9833e2b2d9915ce3ce5c679df08017c80123fa482" Dec 06 14:15:44 crc kubenswrapper[4706]: E1206 14:15:44.110377 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6734078a417bc2c57cba55d9833e2b2d9915ce3ce5c679df08017c80123fa482\": container with ID starting with 6734078a417bc2c57cba55d9833e2b2d9915ce3ce5c679df08017c80123fa482 not found: ID does not exist" containerID="6734078a417bc2c57cba55d9833e2b2d9915ce3ce5c679df08017c80123fa482" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.110415 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6734078a417bc2c57cba55d9833e2b2d9915ce3ce5c679df08017c80123fa482"} err="failed to get container status \"6734078a417bc2c57cba55d9833e2b2d9915ce3ce5c679df08017c80123fa482\": rpc error: code = NotFound desc = could not find container \"6734078a417bc2c57cba55d9833e2b2d9915ce3ce5c679df08017c80123fa482\": container with ID starting with 6734078a417bc2c57cba55d9833e2b2d9915ce3ce5c679df08017c80123fa482 not found: ID does not exist" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.110438 4706 scope.go:117] "RemoveContainer" containerID="98c8c2349e48854c0ef7f4fd6f2a3ac4dbc581f868848b433127270a7dca0cab" Dec 06 14:15:44 crc kubenswrapper[4706]: E1206 14:15:44.110904 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98c8c2349e48854c0ef7f4fd6f2a3ac4dbc581f868848b433127270a7dca0cab\": container with ID starting with 98c8c2349e48854c0ef7f4fd6f2a3ac4dbc581f868848b433127270a7dca0cab not found: ID does not exist" containerID="98c8c2349e48854c0ef7f4fd6f2a3ac4dbc581f868848b433127270a7dca0cab" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.110955 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98c8c2349e48854c0ef7f4fd6f2a3ac4dbc581f868848b433127270a7dca0cab"} err="failed to get container status \"98c8c2349e48854c0ef7f4fd6f2a3ac4dbc581f868848b433127270a7dca0cab\": rpc error: code = NotFound desc = could not find container \"98c8c2349e48854c0ef7f4fd6f2a3ac4dbc581f868848b433127270a7dca0cab\": container with ID starting with 98c8c2349e48854c0ef7f4fd6f2a3ac4dbc581f868848b433127270a7dca0cab not found: ID does not exist" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.389299 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.396548 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.581107 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.581186 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.820934 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-z5f55"] Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.834258 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-z5f55"] Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.844714 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder06a5-account-delete-wrhnv"] Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.851398 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt"] Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.859755 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-06a5-account-create-update-29vqt"] Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.868369 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder06a5-account-delete-wrhnv"] Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.908421 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-create-8kvhz"] Dec 06 14:15:44 crc kubenswrapper[4706]: E1206 14:15:44.908693 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="cinder-volume" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.908714 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="cinder-volume" Dec 06 14:15:44 crc kubenswrapper[4706]: E1206 14:15:44.908733 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b77d62e6-8988-4135-a645-14fcdd3d9c92" containerName="probe" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.908745 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b77d62e6-8988-4135-a645-14fcdd3d9c92" containerName="probe" Dec 06 14:15:44 crc kubenswrapper[4706]: E1206 14:15:44.908765 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b77d62e6-8988-4135-a645-14fcdd3d9c92" containerName="cinder-backup" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.908778 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b77d62e6-8988-4135-a645-14fcdd3d9c92" containerName="cinder-backup" Dec 06 14:15:44 crc kubenswrapper[4706]: E1206 14:15:44.908794 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2316de26-2427-42bf-9d27-f84217c2c343" containerName="probe" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.908801 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="2316de26-2427-42bf-9d27-f84217c2c343" containerName="probe" Dec 06 14:15:44 crc kubenswrapper[4706]: E1206 14:15:44.908814 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2316de26-2427-42bf-9d27-f84217c2c343" containerName="cinder-scheduler" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.908823 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="2316de26-2427-42bf-9d27-f84217c2c343" containerName="cinder-scheduler" Dec 06 14:15:44 crc kubenswrapper[4706]: E1206 14:15:44.908841 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="probe" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.908848 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="probe" Dec 06 14:15:44 crc kubenswrapper[4706]: E1206 14:15:44.908858 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="probe" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.908865 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="probe" Dec 06 14:15:44 crc kubenswrapper[4706]: E1206 14:15:44.908873 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3335ed5-4a44-4d31-ac61-73abbf2a8c18" containerName="mariadb-account-delete" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.908881 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3335ed5-4a44-4d31-ac61-73abbf2a8c18" containerName="mariadb-account-delete" Dec 06 14:15:44 crc kubenswrapper[4706]: E1206 14:15:44.908894 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="cinder-volume" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.908902 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="cinder-volume" Dec 06 14:15:44 crc kubenswrapper[4706]: E1206 14:15:44.908911 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34735cc4-f322-45b7-b2cd-7dd16109952f" containerName="cinder-api-log" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.908926 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="34735cc4-f322-45b7-b2cd-7dd16109952f" containerName="cinder-api-log" Dec 06 14:15:44 crc kubenswrapper[4706]: E1206 14:15:44.908959 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="cinder-volume" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.908967 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="cinder-volume" Dec 06 14:15:44 crc kubenswrapper[4706]: E1206 14:15:44.908981 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34735cc4-f322-45b7-b2cd-7dd16109952f" containerName="cinder-api" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.908988 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="34735cc4-f322-45b7-b2cd-7dd16109952f" containerName="cinder-api" Dec 06 14:15:44 crc kubenswrapper[4706]: E1206 14:15:44.909004 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="probe" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.909012 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="probe" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.909151 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3335ed5-4a44-4d31-ac61-73abbf2a8c18" containerName="mariadb-account-delete" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.909163 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="cinder-volume" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.909174 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="probe" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.909185 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="probe" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.909193 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="34735cc4-f322-45b7-b2cd-7dd16109952f" containerName="cinder-api-log" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.909208 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="2316de26-2427-42bf-9d27-f84217c2c343" containerName="probe" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.909219 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b77d62e6-8988-4135-a645-14fcdd3d9c92" containerName="cinder-backup" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.909231 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="cinder-volume" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.909240 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="cinder-volume" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.909251 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="2316de26-2427-42bf-9d27-f84217c2c343" containerName="cinder-scheduler" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.909262 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b77d62e6-8988-4135-a645-14fcdd3d9c92" containerName="probe" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.909273 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="34735cc4-f322-45b7-b2cd-7dd16109952f" containerName="cinder-api" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.909824 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-8kvhz" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.919694 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-8kvhz"] Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.976239 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk99s\" (UniqueName: \"kubernetes.io/projected/f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99-kube-api-access-kk99s\") pod \"cinder-db-create-8kvhz\" (UID: \"f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99\") " pod="cinder-kuttl-tests/cinder-db-create-8kvhz" Dec 06 14:15:44 crc kubenswrapper[4706]: I1206 14:15:44.976363 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99-operator-scripts\") pod \"cinder-db-create-8kvhz\" (UID: \"f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99\") " pod="cinder-kuttl-tests/cinder-db-create-8kvhz" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.078217 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk99s\" (UniqueName: \"kubernetes.io/projected/f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99-kube-api-access-kk99s\") pod \"cinder-db-create-8kvhz\" (UID: \"f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99\") " pod="cinder-kuttl-tests/cinder-db-create-8kvhz" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.078637 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99-operator-scripts\") pod \"cinder-db-create-8kvhz\" (UID: \"f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99\") " pod="cinder-kuttl-tests/cinder-db-create-8kvhz" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.079308 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99-operator-scripts\") pod \"cinder-db-create-8kvhz\" (UID: \"f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99\") " pod="cinder-kuttl-tests/cinder-db-create-8kvhz" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.096811 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk99s\" (UniqueName: \"kubernetes.io/projected/f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99-kube-api-access-kk99s\") pod \"cinder-db-create-8kvhz\" (UID: \"f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99\") " pod="cinder-kuttl-tests/cinder-db-create-8kvhz" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.114431 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz"] Dec 06 14:15:45 crc kubenswrapper[4706]: E1206 14:15:45.114673 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="probe" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.114685 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="probe" Dec 06 14:15:45 crc kubenswrapper[4706]: E1206 14:15:45.114693 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="cinder-volume" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.114699 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="cinder-volume" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.114802 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="probe" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.114815 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="probe" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.114828 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="49814569-1bcb-4ddd-97db-7115d5ec48d1" containerName="cinder-volume" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.115253 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.116926 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-db-secret" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.136728 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz"] Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.179841 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc4tg\" (UniqueName: \"kubernetes.io/projected/c1a69216-9494-48d4-828a-2a0069314d37-kube-api-access-pc4tg\") pod \"cinder-cf71-account-create-update-svhfz\" (UID: \"c1a69216-9494-48d4-828a-2a0069314d37\") " pod="cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.179918 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1a69216-9494-48d4-828a-2a0069314d37-operator-scripts\") pod \"cinder-cf71-account-create-update-svhfz\" (UID: \"c1a69216-9494-48d4-828a-2a0069314d37\") " pod="cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.225181 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-8kvhz" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.282166 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc4tg\" (UniqueName: \"kubernetes.io/projected/c1a69216-9494-48d4-828a-2a0069314d37-kube-api-access-pc4tg\") pod \"cinder-cf71-account-create-update-svhfz\" (UID: \"c1a69216-9494-48d4-828a-2a0069314d37\") " pod="cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.283766 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1a69216-9494-48d4-828a-2a0069314d37-operator-scripts\") pod \"cinder-cf71-account-create-update-svhfz\" (UID: \"c1a69216-9494-48d4-828a-2a0069314d37\") " pod="cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.282404 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1a69216-9494-48d4-828a-2a0069314d37-operator-scripts\") pod \"cinder-cf71-account-create-update-svhfz\" (UID: \"c1a69216-9494-48d4-828a-2a0069314d37\") " pod="cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.307863 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc4tg\" (UniqueName: \"kubernetes.io/projected/c1a69216-9494-48d4-828a-2a0069314d37-kube-api-access-pc4tg\") pod \"cinder-cf71-account-create-update-svhfz\" (UID: \"c1a69216-9494-48d4-828a-2a0069314d37\") " pod="cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.438398 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.521562 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fe1a084-66ec-4f61-ad67-bda961120ab6" path="/var/lib/kubelet/pods/2fe1a084-66ec-4f61-ad67-bda961120ab6/volumes" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.522702 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34735cc4-f322-45b7-b2cd-7dd16109952f" path="/var/lib/kubelet/pods/34735cc4-f322-45b7-b2cd-7dd16109952f/volumes" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.523614 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b77d62e6-8988-4135-a645-14fcdd3d9c92" path="/var/lib/kubelet/pods/b77d62e6-8988-4135-a645-14fcdd3d9c92/volumes" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.524147 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8e89294-f497-464b-b4d7-3556ce8152a1" path="/var/lib/kubelet/pods/c8e89294-f497-464b-b4d7-3556ce8152a1/volumes" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.525157 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3335ed5-4a44-4d31-ac61-73abbf2a8c18" path="/var/lib/kubelet/pods/d3335ed5-4a44-4d31-ac61-73abbf2a8c18/volumes" Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.525691 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-8kvhz"] Dec 06 14:15:45 crc kubenswrapper[4706]: W1206 14:15:45.527876 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7f0f60f_9f01_4d0e_aaf7_4908d2d0fb99.slice/crio-bbffbd66c91d9808ff24cd5e47ae858fb0e917692a0f1389ff6a45e7bcdaa20c WatchSource:0}: Error finding container bbffbd66c91d9808ff24cd5e47ae858fb0e917692a0f1389ff6a45e7bcdaa20c: Status 404 returned error can't find the container with id bbffbd66c91d9808ff24cd5e47ae858fb0e917692a0f1389ff6a45e7bcdaa20c Dec 06 14:15:45 crc kubenswrapper[4706]: I1206 14:15:45.660146 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz"] Dec 06 14:15:45 crc kubenswrapper[4706]: W1206 14:15:45.661711 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1a69216_9494_48d4_828a_2a0069314d37.slice/crio-46c90b5df54579be5d05c11346788f160774c2eaa2986f00911b325f6ec824cb WatchSource:0}: Error finding container 46c90b5df54579be5d05c11346788f160774c2eaa2986f00911b325f6ec824cb: Status 404 returned error can't find the container with id 46c90b5df54579be5d05c11346788f160774c2eaa2986f00911b325f6ec824cb Dec 06 14:15:46 crc kubenswrapper[4706]: I1206 14:15:46.035466 4706 generic.go:334] "Generic (PLEG): container finished" podID="f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99" containerID="d97ff5efa879357f66a647a69f8a06087361f667a750a4e0db74e62cd38062b6" exitCode=0 Dec 06 14:15:46 crc kubenswrapper[4706]: I1206 14:15:46.035517 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-8kvhz" event={"ID":"f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99","Type":"ContainerDied","Data":"d97ff5efa879357f66a647a69f8a06087361f667a750a4e0db74e62cd38062b6"} Dec 06 14:15:46 crc kubenswrapper[4706]: I1206 14:15:46.035578 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-8kvhz" event={"ID":"f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99","Type":"ContainerStarted","Data":"bbffbd66c91d9808ff24cd5e47ae858fb0e917692a0f1389ff6a45e7bcdaa20c"} Dec 06 14:15:46 crc kubenswrapper[4706]: I1206 14:15:46.037514 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz" event={"ID":"c1a69216-9494-48d4-828a-2a0069314d37","Type":"ContainerStarted","Data":"1fdb9aa35a8f80efeebee1cfff0f12032e090d8f7195dc70cc457a9fac504338"} Dec 06 14:15:46 crc kubenswrapper[4706]: I1206 14:15:46.037547 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz" event={"ID":"c1a69216-9494-48d4-828a-2a0069314d37","Type":"ContainerStarted","Data":"46c90b5df54579be5d05c11346788f160774c2eaa2986f00911b325f6ec824cb"} Dec 06 14:15:46 crc kubenswrapper[4706]: I1206 14:15:46.079691 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz" podStartSLOduration=1.079667389 podStartE2EDuration="1.079667389s" podCreationTimestamp="2025-12-06 14:15:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:15:46.073681361 +0000 UTC m=+1188.909474258" watchObservedRunningTime="2025-12-06 14:15:46.079667389 +0000 UTC m=+1188.915460286" Dec 06 14:15:47 crc kubenswrapper[4706]: I1206 14:15:47.056351 4706 generic.go:334] "Generic (PLEG): container finished" podID="c1a69216-9494-48d4-828a-2a0069314d37" containerID="1fdb9aa35a8f80efeebee1cfff0f12032e090d8f7195dc70cc457a9fac504338" exitCode=0 Dec 06 14:15:47 crc kubenswrapper[4706]: I1206 14:15:47.056612 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz" event={"ID":"c1a69216-9494-48d4-828a-2a0069314d37","Type":"ContainerDied","Data":"1fdb9aa35a8f80efeebee1cfff0f12032e090d8f7195dc70cc457a9fac504338"} Dec 06 14:15:47 crc kubenswrapper[4706]: I1206 14:15:47.386748 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-8kvhz" Dec 06 14:15:47 crc kubenswrapper[4706]: I1206 14:15:47.420369 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kk99s\" (UniqueName: \"kubernetes.io/projected/f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99-kube-api-access-kk99s\") pod \"f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99\" (UID: \"f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99\") " Dec 06 14:15:47 crc kubenswrapper[4706]: I1206 14:15:47.420440 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99-operator-scripts\") pod \"f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99\" (UID: \"f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99\") " Dec 06 14:15:47 crc kubenswrapper[4706]: I1206 14:15:47.421772 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99" (UID: "f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:15:47 crc kubenswrapper[4706]: I1206 14:15:47.429399 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99-kube-api-access-kk99s" (OuterVolumeSpecName: "kube-api-access-kk99s") pod "f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99" (UID: "f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99"). InnerVolumeSpecName "kube-api-access-kk99s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:15:47 crc kubenswrapper[4706]: I1206 14:15:47.523969 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kk99s\" (UniqueName: \"kubernetes.io/projected/f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99-kube-api-access-kk99s\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:47 crc kubenswrapper[4706]: I1206 14:15:47.524057 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:48 crc kubenswrapper[4706]: I1206 14:15:48.066283 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-8kvhz" Dec 06 14:15:48 crc kubenswrapper[4706]: I1206 14:15:48.066278 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-8kvhz" event={"ID":"f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99","Type":"ContainerDied","Data":"bbffbd66c91d9808ff24cd5e47ae858fb0e917692a0f1389ff6a45e7bcdaa20c"} Dec 06 14:15:48 crc kubenswrapper[4706]: I1206 14:15:48.066535 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbffbd66c91d9808ff24cd5e47ae858fb0e917692a0f1389ff6a45e7bcdaa20c" Dec 06 14:15:48 crc kubenswrapper[4706]: I1206 14:15:48.396962 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz" Dec 06 14:15:48 crc kubenswrapper[4706]: I1206 14:15:48.438792 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1a69216-9494-48d4-828a-2a0069314d37-operator-scripts\") pod \"c1a69216-9494-48d4-828a-2a0069314d37\" (UID: \"c1a69216-9494-48d4-828a-2a0069314d37\") " Dec 06 14:15:48 crc kubenswrapper[4706]: I1206 14:15:48.438935 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pc4tg\" (UniqueName: \"kubernetes.io/projected/c1a69216-9494-48d4-828a-2a0069314d37-kube-api-access-pc4tg\") pod \"c1a69216-9494-48d4-828a-2a0069314d37\" (UID: \"c1a69216-9494-48d4-828a-2a0069314d37\") " Dec 06 14:15:48 crc kubenswrapper[4706]: I1206 14:15:48.439524 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1a69216-9494-48d4-828a-2a0069314d37-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c1a69216-9494-48d4-828a-2a0069314d37" (UID: "c1a69216-9494-48d4-828a-2a0069314d37"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:15:48 crc kubenswrapper[4706]: I1206 14:15:48.443071 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1a69216-9494-48d4-828a-2a0069314d37-kube-api-access-pc4tg" (OuterVolumeSpecName: "kube-api-access-pc4tg") pod "c1a69216-9494-48d4-828a-2a0069314d37" (UID: "c1a69216-9494-48d4-828a-2a0069314d37"). InnerVolumeSpecName "kube-api-access-pc4tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:15:48 crc kubenswrapper[4706]: I1206 14:15:48.541508 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pc4tg\" (UniqueName: \"kubernetes.io/projected/c1a69216-9494-48d4-828a-2a0069314d37-kube-api-access-pc4tg\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:48 crc kubenswrapper[4706]: I1206 14:15:48.541539 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1a69216-9494-48d4-828a-2a0069314d37-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:49 crc kubenswrapper[4706]: I1206 14:15:49.076818 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz" event={"ID":"c1a69216-9494-48d4-828a-2a0069314d37","Type":"ContainerDied","Data":"46c90b5df54579be5d05c11346788f160774c2eaa2986f00911b325f6ec824cb"} Dec 06 14:15:49 crc kubenswrapper[4706]: I1206 14:15:49.076870 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46c90b5df54579be5d05c11346788f160774c2eaa2986f00911b325f6ec824cb" Dec 06 14:15:49 crc kubenswrapper[4706]: I1206 14:15:49.076920 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.294702 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-bv2sh"] Dec 06 14:15:50 crc kubenswrapper[4706]: E1206 14:15:50.295317 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99" containerName="mariadb-database-create" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.295349 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99" containerName="mariadb-database-create" Dec 06 14:15:50 crc kubenswrapper[4706]: E1206 14:15:50.295365 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1a69216-9494-48d4-828a-2a0069314d37" containerName="mariadb-account-create-update" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.295377 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1a69216-9494-48d4-828a-2a0069314d37" containerName="mariadb-account-create-update" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.295526 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99" containerName="mariadb-database-create" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.295538 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1a69216-9494-48d4-828a-2a0069314d37" containerName="mariadb-account-create-update" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.296177 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.302129 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"combined-ca-bundle" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.302604 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.303086 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-zb5st" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.303117 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.318730 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-bv2sh"] Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.371048 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cad87657-2720-489c-b6c1-386b6e967c47-etc-machine-id\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.371112 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-combined-ca-bundle\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.371158 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xpmg\" (UniqueName: \"kubernetes.io/projected/cad87657-2720-489c-b6c1-386b6e967c47-kube-api-access-6xpmg\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.371206 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-scripts\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.371363 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-config-data\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.371423 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-db-sync-config-data\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.473316 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cad87657-2720-489c-b6c1-386b6e967c47-etc-machine-id\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.473390 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-combined-ca-bundle\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.473417 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xpmg\" (UniqueName: \"kubernetes.io/projected/cad87657-2720-489c-b6c1-386b6e967c47-kube-api-access-6xpmg\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.473448 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-scripts\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.473473 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-config-data\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.473473 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cad87657-2720-489c-b6c1-386b6e967c47-etc-machine-id\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.473489 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-db-sync-config-data\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.479972 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-db-sync-config-data\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.479996 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-scripts\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.480075 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-combined-ca-bundle\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.481069 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-config-data\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.498120 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xpmg\" (UniqueName: \"kubernetes.io/projected/cad87657-2720-489c-b6c1-386b6e967c47-kube-api-access-6xpmg\") pod \"cinder-db-sync-bv2sh\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:50 crc kubenswrapper[4706]: I1206 14:15:50.627600 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:51 crc kubenswrapper[4706]: I1206 14:15:51.112020 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-bv2sh"] Dec 06 14:15:52 crc kubenswrapper[4706]: I1206 14:15:52.104688 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" event={"ID":"cad87657-2720-489c-b6c1-386b6e967c47","Type":"ContainerStarted","Data":"8e9515fa0aa616d09ed6f6937c8190ed7aaff8a54888f168263c19f6ed4bfc5e"} Dec 06 14:15:52 crc kubenswrapper[4706]: I1206 14:15:52.105024 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" event={"ID":"cad87657-2720-489c-b6c1-386b6e967c47","Type":"ContainerStarted","Data":"609e3eee9072b6426b1e7863592b35778e5f1026ec31399072218cb89a00b606"} Dec 06 14:15:52 crc kubenswrapper[4706]: I1206 14:15:52.123799 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" podStartSLOduration=2.123780833 podStartE2EDuration="2.123780833s" podCreationTimestamp="2025-12-06 14:15:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:15:52.120436514 +0000 UTC m=+1194.956229411" watchObservedRunningTime="2025-12-06 14:15:52.123780833 +0000 UTC m=+1194.959573710" Dec 06 14:15:54 crc kubenswrapper[4706]: I1206 14:15:54.126614 4706 generic.go:334] "Generic (PLEG): container finished" podID="cad87657-2720-489c-b6c1-386b6e967c47" containerID="8e9515fa0aa616d09ed6f6937c8190ed7aaff8a54888f168263c19f6ed4bfc5e" exitCode=0 Dec 06 14:15:54 crc kubenswrapper[4706]: I1206 14:15:54.126654 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" event={"ID":"cad87657-2720-489c-b6c1-386b6e967c47","Type":"ContainerDied","Data":"8e9515fa0aa616d09ed6f6937c8190ed7aaff8a54888f168263c19f6ed4bfc5e"} Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.519763 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.549707 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cad87657-2720-489c-b6c1-386b6e967c47-etc-machine-id\") pod \"cad87657-2720-489c-b6c1-386b6e967c47\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.550010 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-db-sync-config-data\") pod \"cad87657-2720-489c-b6c1-386b6e967c47\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.550053 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-scripts\") pod \"cad87657-2720-489c-b6c1-386b6e967c47\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.550234 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-combined-ca-bundle\") pod \"cad87657-2720-489c-b6c1-386b6e967c47\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.550291 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-config-data\") pod \"cad87657-2720-489c-b6c1-386b6e967c47\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.550355 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xpmg\" (UniqueName: \"kubernetes.io/projected/cad87657-2720-489c-b6c1-386b6e967c47-kube-api-access-6xpmg\") pod \"cad87657-2720-489c-b6c1-386b6e967c47\" (UID: \"cad87657-2720-489c-b6c1-386b6e967c47\") " Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.549845 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cad87657-2720-489c-b6c1-386b6e967c47-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cad87657-2720-489c-b6c1-386b6e967c47" (UID: "cad87657-2720-489c-b6c1-386b6e967c47"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.550891 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cad87657-2720-489c-b6c1-386b6e967c47-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.555322 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-scripts" (OuterVolumeSpecName: "scripts") pod "cad87657-2720-489c-b6c1-386b6e967c47" (UID: "cad87657-2720-489c-b6c1-386b6e967c47"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.558054 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cad87657-2720-489c-b6c1-386b6e967c47-kube-api-access-6xpmg" (OuterVolumeSpecName: "kube-api-access-6xpmg") pod "cad87657-2720-489c-b6c1-386b6e967c47" (UID: "cad87657-2720-489c-b6c1-386b6e967c47"). InnerVolumeSpecName "kube-api-access-6xpmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.561465 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "cad87657-2720-489c-b6c1-386b6e967c47" (UID: "cad87657-2720-489c-b6c1-386b6e967c47"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.584321 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cad87657-2720-489c-b6c1-386b6e967c47" (UID: "cad87657-2720-489c-b6c1-386b6e967c47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.601855 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-config-data" (OuterVolumeSpecName: "config-data") pod "cad87657-2720-489c-b6c1-386b6e967c47" (UID: "cad87657-2720-489c-b6c1-386b6e967c47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.652456 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.652496 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.652511 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.652525 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xpmg\" (UniqueName: \"kubernetes.io/projected/cad87657-2720-489c-b6c1-386b6e967c47-kube-api-access-6xpmg\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:55 crc kubenswrapper[4706]: I1206 14:15:55.652538 4706 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cad87657-2720-489c-b6c1-386b6e967c47-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.151034 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" event={"ID":"cad87657-2720-489c-b6c1-386b6e967c47","Type":"ContainerDied","Data":"609e3eee9072b6426b1e7863592b35778e5f1026ec31399072218cb89a00b606"} Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.151092 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="609e3eee9072b6426b1e7863592b35778e5f1026ec31399072218cb89a00b606" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.151135 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-bv2sh" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.393868 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 06 14:15:56 crc kubenswrapper[4706]: E1206 14:15:56.394151 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cad87657-2720-489c-b6c1-386b6e967c47" containerName="cinder-db-sync" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.394165 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="cad87657-2720-489c-b6c1-386b6e967c47" containerName="cinder-db-sync" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.394568 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="cad87657-2720-489c-b6c1-386b6e967c47" containerName="cinder-db-sync" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.395637 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.398027 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"combined-ca-bundle" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.398354 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-backup-config-data" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.399812 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.400618 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-zb5st" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.403894 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.430862 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.431983 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.434209 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scheduler-config-data" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461097 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc55b\" (UniqueName: \"kubernetes.io/projected/059b6eae-b9fa-43fd-92a4-f61be1796941-kube-api-access-nc55b\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461144 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-scripts\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461174 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-lib-modules\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461196 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461214 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-run\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461231 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461260 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/059b6eae-b9fa-43fd-92a4-f61be1796941-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461374 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461405 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461430 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-dev\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461449 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-config-data-custom\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461466 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461501 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-config-data\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461521 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqdz5\" (UniqueName: \"kubernetes.io/projected/f10bc859-5282-4d77-8287-823c2f480bc1-kube-api-access-xqdz5\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461537 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-config-data\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461643 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461686 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-nvme\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461723 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461744 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-sys\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461781 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.461800 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-scripts\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.485543 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.510875 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.512213 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.518584 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-volume-volume1-config-data" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.519704 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.563628 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.563690 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-nvme\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.563737 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.563774 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.563785 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-nvme\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.563797 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-sys\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.563857 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.563883 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.563902 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-sys\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.563926 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.563919 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.563952 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfbvk\" (UniqueName: \"kubernetes.io/projected/b3ea9d97-694e-43f8-8747-64c34bdab8dd-kube-api-access-dfbvk\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564046 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-scripts\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564081 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564105 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564124 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564157 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-run\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564183 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc55b\" (UniqueName: \"kubernetes.io/projected/059b6eae-b9fa-43fd-92a4-f61be1796941-kube-api-access-nc55b\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564212 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564234 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-lib-modules\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564255 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564274 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-scripts\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564299 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564336 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-run\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564357 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564404 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/059b6eae-b9fa-43fd-92a4-f61be1796941-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564432 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564452 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564479 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564490 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-lib-modules\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564503 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-dev\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564533 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-dev\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564545 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-config-data-custom\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564568 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564582 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/059b6eae-b9fa-43fd-92a4-f61be1796941-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564635 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.563762 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564884 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-config-data\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564907 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564927 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-run\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.564931 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.565455 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-sys\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.565482 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqdz5\" (UniqueName: \"kubernetes.io/projected/f10bc859-5282-4d77-8287-823c2f480bc1-kube-api-access-xqdz5\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.565513 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-dev\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.565763 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-config-data\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.565925 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.565993 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.572488 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.574810 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-config-data-custom\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.574988 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.591606 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-config-data\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.593118 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.594841 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.595798 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-scripts\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.611449 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-config-data\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.613938 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc55b\" (UniqueName: \"kubernetes.io/projected/059b6eae-b9fa-43fd-92a4-f61be1796941-kube-api-access-nc55b\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.627926 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqdz5\" (UniqueName: \"kubernetes.io/projected/f10bc859-5282-4d77-8287-823c2f480bc1-kube-api-access-xqdz5\") pod \"cinder-backup-0\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.663423 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-scripts\") pod \"cinder-scheduler-0\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668137 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668194 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-dev\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668219 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668244 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668281 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668333 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668353 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-sys\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668376 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668396 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfbvk\" (UniqueName: \"kubernetes.io/projected/b3ea9d97-694e-43f8-8747-64c34bdab8dd-kube-api-access-dfbvk\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668417 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668435 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668455 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668477 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-run\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668509 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668527 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668642 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668688 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.668716 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-dev\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.669776 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.669853 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.670505 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.670564 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.670609 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.670643 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-sys\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.670765 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-run\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.682933 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.683753 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.684249 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.694266 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.700880 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfbvk\" (UniqueName: \"kubernetes.io/projected/b3ea9d97-694e-43f8-8747-64c34bdab8dd-kube-api-access-dfbvk\") pod \"cinder-volume-volume1-0\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.715661 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.742970 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.744326 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.746460 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-api-config-data" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.746557 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.747608 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cert-cinder-internal-svc" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.748441 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.749112 4706 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cert-cinder-public-svc" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.771034 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5375678a-f779-46ba-bcf4-71932e9ac6c5-logs\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.771082 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-config-data-custom\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.771150 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkjrk\" (UniqueName: \"kubernetes.io/projected/5375678a-f779-46ba-bcf4-71932e9ac6c5-kube-api-access-dkjrk\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.771180 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-config-data\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.771390 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.771443 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5375678a-f779-46ba-bcf4-71932e9ac6c5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.771607 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.771680 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-scripts\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.771725 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.827429 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.873690 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-config-data\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.873817 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.873844 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5375678a-f779-46ba-bcf4-71932e9ac6c5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.874717 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.874753 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-scripts\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.874779 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.874807 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5375678a-f779-46ba-bcf4-71932e9ac6c5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.874836 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5375678a-f779-46ba-bcf4-71932e9ac6c5-logs\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.874919 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-config-data-custom\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.874980 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkjrk\" (UniqueName: \"kubernetes.io/projected/5375678a-f779-46ba-bcf4-71932e9ac6c5-kube-api-access-dkjrk\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.875175 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5375678a-f779-46ba-bcf4-71932e9ac6c5-logs\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.878040 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.878845 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-config-data-custom\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.878907 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.879395 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-config-data\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.881892 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-scripts\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.884691 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:56 crc kubenswrapper[4706]: I1206 14:15:56.897472 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkjrk\" (UniqueName: \"kubernetes.io/projected/5375678a-f779-46ba-bcf4-71932e9ac6c5-kube-api-access-dkjrk\") pod \"cinder-api-0\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:57 crc kubenswrapper[4706]: I1206 14:15:57.060457 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:57 crc kubenswrapper[4706]: I1206 14:15:57.098020 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 06 14:15:57 crc kubenswrapper[4706]: I1206 14:15:57.176573 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"b3ea9d97-694e-43f8-8747-64c34bdab8dd","Type":"ContainerStarted","Data":"66789396af9bc12e94585b205b3eb9fc307afc81380f9624cdf46a8a86956650"} Dec 06 14:15:57 crc kubenswrapper[4706]: I1206 14:15:57.190240 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 06 14:15:57 crc kubenswrapper[4706]: I1206 14:15:57.252526 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 06 14:15:57 crc kubenswrapper[4706]: W1206 14:15:57.265196 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod059b6eae_b9fa_43fd_92a4_f61be1796941.slice/crio-a3196368460110a6ca17e240b421e254d95e70eaf8ad91b3ed4055e738651ece WatchSource:0}: Error finding container a3196368460110a6ca17e240b421e254d95e70eaf8ad91b3ed4055e738651ece: Status 404 returned error can't find the container with id a3196368460110a6ca17e240b421e254d95e70eaf8ad91b3ed4055e738651ece Dec 06 14:15:57 crc kubenswrapper[4706]: I1206 14:15:57.566252 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:15:58 crc kubenswrapper[4706]: I1206 14:15:58.190652 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"b3ea9d97-694e-43f8-8747-64c34bdab8dd","Type":"ContainerStarted","Data":"b4df2faaae1d87f67c4e865f411c79d7245502c38137722bbfd7ca4967b22137"} Dec 06 14:15:58 crc kubenswrapper[4706]: I1206 14:15:58.191176 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"b3ea9d97-694e-43f8-8747-64c34bdab8dd","Type":"ContainerStarted","Data":"9dd17b54d2c3985f57ba5f12923110963ab6c35567202d409bbdd856f16e21f4"} Dec 06 14:15:58 crc kubenswrapper[4706]: I1206 14:15:58.195816 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"059b6eae-b9fa-43fd-92a4-f61be1796941","Type":"ContainerStarted","Data":"b24aaf8d7884a0230176a110498ef30b0704f3d3b9162a0773117e03b098fd9c"} Dec 06 14:15:58 crc kubenswrapper[4706]: I1206 14:15:58.195862 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"059b6eae-b9fa-43fd-92a4-f61be1796941","Type":"ContainerStarted","Data":"a3196368460110a6ca17e240b421e254d95e70eaf8ad91b3ed4055e738651ece"} Dec 06 14:15:58 crc kubenswrapper[4706]: I1206 14:15:58.202912 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"5375678a-f779-46ba-bcf4-71932e9ac6c5","Type":"ContainerStarted","Data":"8c5bfcbedb802b8f2a803098e5de887d06b1903a1266ca736ab852c4931fcf0b"} Dec 06 14:15:58 crc kubenswrapper[4706]: I1206 14:15:58.206342 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"f10bc859-5282-4d77-8287-823c2f480bc1","Type":"ContainerStarted","Data":"f1875d1e195ab51c83a5ab7dc59f64832d00d51b571cec82c2a8bab8833ac405"} Dec 06 14:15:58 crc kubenswrapper[4706]: I1206 14:15:58.206437 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"f10bc859-5282-4d77-8287-823c2f480bc1","Type":"ContainerStarted","Data":"103ec7c6f4792616852ef5c8055c1871f9d15d6bdf195f49a46f5ce642f0d340"} Dec 06 14:15:58 crc kubenswrapper[4706]: I1206 14:15:58.206451 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"f10bc859-5282-4d77-8287-823c2f480bc1","Type":"ContainerStarted","Data":"6abe933ab9e15b5cd35821769c8bc23a562ec307dbbd0c6f09b79f66e04ff01a"} Dec 06 14:15:58 crc kubenswrapper[4706]: I1206 14:15:58.225676 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podStartSLOduration=2.225653008 podStartE2EDuration="2.225653008s" podCreationTimestamp="2025-12-06 14:15:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:15:58.213615199 +0000 UTC m=+1201.049408066" watchObservedRunningTime="2025-12-06 14:15:58.225653008 +0000 UTC m=+1201.061445875" Dec 06 14:15:59 crc kubenswrapper[4706]: I1206 14:15:59.214982 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"059b6eae-b9fa-43fd-92a4-f61be1796941","Type":"ContainerStarted","Data":"c85226ca8172c2f9b2711df6efb2637a4b4965da96001b821bbb8e2b00799052"} Dec 06 14:15:59 crc kubenswrapper[4706]: I1206 14:15:59.218136 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"5375678a-f779-46ba-bcf4-71932e9ac6c5","Type":"ContainerStarted","Data":"34fb37f23daecfa827f3d223a2d3f2a11c676aca2295b0bcba9329c001b0dbe3"} Dec 06 14:15:59 crc kubenswrapper[4706]: I1206 14:15:59.218169 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"5375678a-f779-46ba-bcf4-71932e9ac6c5","Type":"ContainerStarted","Data":"66815d90ff92c4f1709453e09ae9af94254ea4357ac10399dfb261a4a975901c"} Dec 06 14:15:59 crc kubenswrapper[4706]: I1206 14:15:59.218202 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:15:59 crc kubenswrapper[4706]: I1206 14:15:59.236110 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-0" podStartSLOduration=3.236090933 podStartE2EDuration="3.236090933s" podCreationTimestamp="2025-12-06 14:15:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:15:58.256552797 +0000 UTC m=+1201.092345654" watchObservedRunningTime="2025-12-06 14:15:59.236090933 +0000 UTC m=+1202.071883790" Dec 06 14:15:59 crc kubenswrapper[4706]: I1206 14:15:59.266933 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-0" podStartSLOduration=3.26690995 podStartE2EDuration="3.26690995s" podCreationTimestamp="2025-12-06 14:15:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:15:59.241292411 +0000 UTC m=+1202.077085268" watchObservedRunningTime="2025-12-06 14:15:59.26690995 +0000 UTC m=+1202.102702827" Dec 06 14:15:59 crc kubenswrapper[4706]: I1206 14:15:59.271911 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-0" podStartSLOduration=3.2718938619999998 podStartE2EDuration="3.271893862s" podCreationTimestamp="2025-12-06 14:15:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:15:59.267935657 +0000 UTC m=+1202.103728514" watchObservedRunningTime="2025-12-06 14:15:59.271893862 +0000 UTC m=+1202.107686719" Dec 06 14:16:00 crc kubenswrapper[4706]: I1206 14:16:00.237619 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"b3ea9d97-694e-43f8-8747-64c34bdab8dd","Type":"ContainerDied","Data":"b4df2faaae1d87f67c4e865f411c79d7245502c38137722bbfd7ca4967b22137"} Dec 06 14:16:00 crc kubenswrapper[4706]: I1206 14:16:00.239246 4706 scope.go:117] "RemoveContainer" containerID="9dd17b54d2c3985f57ba5f12923110963ab6c35567202d409bbdd856f16e21f4" Dec 06 14:16:00 crc kubenswrapper[4706]: I1206 14:16:00.239341 4706 scope.go:117] "RemoveContainer" containerID="b4df2faaae1d87f67c4e865f411c79d7245502c38137722bbfd7ca4967b22137" Dec 06 14:16:00 crc kubenswrapper[4706]: I1206 14:16:00.237514 4706 generic.go:334] "Generic (PLEG): container finished" podID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" containerID="b4df2faaae1d87f67c4e865f411c79d7245502c38137722bbfd7ca4967b22137" exitCode=1 Dec 06 14:16:00 crc kubenswrapper[4706]: I1206 14:16:00.240167 4706 generic.go:334] "Generic (PLEG): container finished" podID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" containerID="9dd17b54d2c3985f57ba5f12923110963ab6c35567202d409bbdd856f16e21f4" exitCode=1 Dec 06 14:16:00 crc kubenswrapper[4706]: I1206 14:16:00.240205 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"b3ea9d97-694e-43f8-8747-64c34bdab8dd","Type":"ContainerDied","Data":"9dd17b54d2c3985f57ba5f12923110963ab6c35567202d409bbdd856f16e21f4"} Dec 06 14:16:01 crc kubenswrapper[4706]: I1206 14:16:01.254065 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"b3ea9d97-694e-43f8-8747-64c34bdab8dd","Type":"ContainerStarted","Data":"16afbb7105a12e7852686ed389c4ae21ed4ba22f8eeefe72c007f7bc7acafc62"} Dec 06 14:16:01 crc kubenswrapper[4706]: I1206 14:16:01.254592 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"b3ea9d97-694e-43f8-8747-64c34bdab8dd","Type":"ContainerStarted","Data":"15687f42b4d952dc619dc822b1a51e51de9aa12b0989f3dfd215c26253896e0f"} Dec 06 14:16:01 crc kubenswrapper[4706]: I1206 14:16:01.716124 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:16:01 crc kubenswrapper[4706]: I1206 14:16:01.747053 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:16:01 crc kubenswrapper[4706]: I1206 14:16:01.828006 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:16:02 crc kubenswrapper[4706]: I1206 14:16:02.266045 4706 generic.go:334] "Generic (PLEG): container finished" podID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" containerID="15687f42b4d952dc619dc822b1a51e51de9aa12b0989f3dfd215c26253896e0f" exitCode=1 Dec 06 14:16:02 crc kubenswrapper[4706]: I1206 14:16:02.266088 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"b3ea9d97-694e-43f8-8747-64c34bdab8dd","Type":"ContainerDied","Data":"15687f42b4d952dc619dc822b1a51e51de9aa12b0989f3dfd215c26253896e0f"} Dec 06 14:16:02 crc kubenswrapper[4706]: I1206 14:16:02.266123 4706 scope.go:117] "RemoveContainer" containerID="9dd17b54d2c3985f57ba5f12923110963ab6c35567202d409bbdd856f16e21f4" Dec 06 14:16:02 crc kubenswrapper[4706]: I1206 14:16:02.266533 4706 scope.go:117] "RemoveContainer" containerID="15687f42b4d952dc619dc822b1a51e51de9aa12b0989f3dfd215c26253896e0f" Dec 06 14:16:02 crc kubenswrapper[4706]: E1206 14:16:02.266858 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(b3ea9d97-694e-43f8-8747-64c34bdab8dd)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" Dec 06 14:16:02 crc kubenswrapper[4706]: I1206 14:16:02.828198 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:16:03 crc kubenswrapper[4706]: I1206 14:16:03.275408 4706 generic.go:334] "Generic (PLEG): container finished" podID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" containerID="16afbb7105a12e7852686ed389c4ae21ed4ba22f8eeefe72c007f7bc7acafc62" exitCode=1 Dec 06 14:16:03 crc kubenswrapper[4706]: I1206 14:16:03.275456 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"b3ea9d97-694e-43f8-8747-64c34bdab8dd","Type":"ContainerDied","Data":"16afbb7105a12e7852686ed389c4ae21ed4ba22f8eeefe72c007f7bc7acafc62"} Dec 06 14:16:03 crc kubenswrapper[4706]: I1206 14:16:03.275496 4706 scope.go:117] "RemoveContainer" containerID="b4df2faaae1d87f67c4e865f411c79d7245502c38137722bbfd7ca4967b22137" Dec 06 14:16:03 crc kubenswrapper[4706]: I1206 14:16:03.276168 4706 scope.go:117] "RemoveContainer" containerID="15687f42b4d952dc619dc822b1a51e51de9aa12b0989f3dfd215c26253896e0f" Dec 06 14:16:03 crc kubenswrapper[4706]: I1206 14:16:03.276196 4706 scope.go:117] "RemoveContainer" containerID="16afbb7105a12e7852686ed389c4ae21ed4ba22f8eeefe72c007f7bc7acafc62" Dec 06 14:16:03 crc kubenswrapper[4706]: E1206 14:16:03.276628 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(b3ea9d97-694e-43f8-8747-64c34bdab8dd)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(b3ea9d97-694e-43f8-8747-64c34bdab8dd)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" Dec 06 14:16:04 crc kubenswrapper[4706]: I1206 14:16:04.287717 4706 scope.go:117] "RemoveContainer" containerID="15687f42b4d952dc619dc822b1a51e51de9aa12b0989f3dfd215c26253896e0f" Dec 06 14:16:04 crc kubenswrapper[4706]: I1206 14:16:04.287761 4706 scope.go:117] "RemoveContainer" containerID="16afbb7105a12e7852686ed389c4ae21ed4ba22f8eeefe72c007f7bc7acafc62" Dec 06 14:16:04 crc kubenswrapper[4706]: E1206 14:16:04.288063 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(b3ea9d97-694e-43f8-8747-64c34bdab8dd)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(b3ea9d97-694e-43f8-8747-64c34bdab8dd)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" Dec 06 14:16:06 crc kubenswrapper[4706]: I1206 14:16:06.828624 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:16:06 crc kubenswrapper[4706]: I1206 14:16:06.829456 4706 scope.go:117] "RemoveContainer" containerID="15687f42b4d952dc619dc822b1a51e51de9aa12b0989f3dfd215c26253896e0f" Dec 06 14:16:06 crc kubenswrapper[4706]: I1206 14:16:06.829472 4706 scope.go:117] "RemoveContainer" containerID="16afbb7105a12e7852686ed389c4ae21ed4ba22f8eeefe72c007f7bc7acafc62" Dec 06 14:16:06 crc kubenswrapper[4706]: E1206 14:16:06.829734 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(b3ea9d97-694e-43f8-8747-64c34bdab8dd)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(b3ea9d97-694e-43f8-8747-64c34bdab8dd)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" Dec 06 14:16:06 crc kubenswrapper[4706]: I1206 14:16:06.913919 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:16:06 crc kubenswrapper[4706]: I1206 14:16:06.997823 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:16:08 crc kubenswrapper[4706]: I1206 14:16:08.745983 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.399693 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-bv2sh"] Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.411239 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-bv2sh"] Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.419835 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cindercf71-account-delete-j99bm"] Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.420656 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cindercf71-account-delete-j99bm" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.442847 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.443106 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="059b6eae-b9fa-43fd-92a4-f61be1796941" containerName="cinder-scheduler" containerID="cri-o://b24aaf8d7884a0230176a110498ef30b0704f3d3b9162a0773117e03b098fd9c" gracePeriod=30 Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.443237 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="059b6eae-b9fa-43fd-92a4-f61be1796941" containerName="probe" containerID="cri-o://c85226ca8172c2f9b2711df6efb2637a4b4965da96001b821bbb8e2b00799052" gracePeriod=30 Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.461171 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cindercf71-account-delete-j99bm"] Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.476761 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.477282 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="5375678a-f779-46ba-bcf4-71932e9ac6c5" containerName="cinder-api-log" containerID="cri-o://66815d90ff92c4f1709453e09ae9af94254ea4357ac10399dfb261a4a975901c" gracePeriod=30 Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.477765 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="5375678a-f779-46ba-bcf4-71932e9ac6c5" containerName="cinder-api" containerID="cri-o://34fb37f23daecfa827f3d223a2d3f2a11c676aca2295b0bcba9329c001b0dbe3" gracePeriod=30 Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.482874 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="cinder-kuttl-tests/cinder-api-0" podUID="5375678a-f779-46ba-bcf4-71932e9ac6c5" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.108:8776/healthcheck\": EOF" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.497793 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f400a2f-eb0c-4a3d-a1a3-463b437b33d6-operator-scripts\") pod \"cindercf71-account-delete-j99bm\" (UID: \"5f400a2f-eb0c-4a3d-a1a3-463b437b33d6\") " pod="cinder-kuttl-tests/cindercf71-account-delete-j99bm" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.498053 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdkhd\" (UniqueName: \"kubernetes.io/projected/5f400a2f-eb0c-4a3d-a1a3-463b437b33d6-kube-api-access-hdkhd\") pod \"cindercf71-account-delete-j99bm\" (UID: \"5f400a2f-eb0c-4a3d-a1a3-463b437b33d6\") " pod="cinder-kuttl-tests/cindercf71-account-delete-j99bm" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.510822 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.522424 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.522721 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="f10bc859-5282-4d77-8287-823c2f480bc1" containerName="cinder-backup" containerID="cri-o://103ec7c6f4792616852ef5c8055c1871f9d15d6bdf195f49a46f5ce642f0d340" gracePeriod=30 Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.522912 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="f10bc859-5282-4d77-8287-823c2f480bc1" containerName="probe" containerID="cri-o://f1875d1e195ab51c83a5ab7dc59f64832d00d51b571cec82c2a8bab8833ac405" gracePeriod=30 Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.600044 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdkhd\" (UniqueName: \"kubernetes.io/projected/5f400a2f-eb0c-4a3d-a1a3-463b437b33d6-kube-api-access-hdkhd\") pod \"cindercf71-account-delete-j99bm\" (UID: \"5f400a2f-eb0c-4a3d-a1a3-463b437b33d6\") " pod="cinder-kuttl-tests/cindercf71-account-delete-j99bm" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.600197 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f400a2f-eb0c-4a3d-a1a3-463b437b33d6-operator-scripts\") pod \"cindercf71-account-delete-j99bm\" (UID: \"5f400a2f-eb0c-4a3d-a1a3-463b437b33d6\") " pod="cinder-kuttl-tests/cindercf71-account-delete-j99bm" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.601066 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f400a2f-eb0c-4a3d-a1a3-463b437b33d6-operator-scripts\") pod \"cindercf71-account-delete-j99bm\" (UID: \"5f400a2f-eb0c-4a3d-a1a3-463b437b33d6\") " pod="cinder-kuttl-tests/cindercf71-account-delete-j99bm" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.642213 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdkhd\" (UniqueName: \"kubernetes.io/projected/5f400a2f-eb0c-4a3d-a1a3-463b437b33d6-kube-api-access-hdkhd\") pod \"cindercf71-account-delete-j99bm\" (UID: \"5f400a2f-eb0c-4a3d-a1a3-463b437b33d6\") " pod="cinder-kuttl-tests/cindercf71-account-delete-j99bm" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.757984 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cindercf71-account-delete-j99bm" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.819966 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.903261 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-nvme\") pod \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.903320 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-config-data-custom\") pod \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.903351 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-locks-cinder\") pod \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.903401 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-run\") pod \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.903439 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfbvk\" (UniqueName: \"kubernetes.io/projected/b3ea9d97-694e-43f8-8747-64c34bdab8dd-kube-api-access-dfbvk\") pod \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.903434 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "b3ea9d97-694e-43f8-8747-64c34bdab8dd" (UID: "b3ea9d97-694e-43f8-8747-64c34bdab8dd"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.903469 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-dev\") pod \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.903498 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-iscsi\") pod \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.903521 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-machine-id\") pod \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.903519 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "b3ea9d97-694e-43f8-8747-64c34bdab8dd" (UID: "b3ea9d97-694e-43f8-8747-64c34bdab8dd"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.903543 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-scripts\") pod \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.904333 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-lib-modules\") pod \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.904399 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-locks-brick\") pod \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.908012 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-sys\") pod \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.904339 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-dev" (OuterVolumeSpecName: "dev") pod "b3ea9d97-694e-43f8-8747-64c34bdab8dd" (UID: "b3ea9d97-694e-43f8-8747-64c34bdab8dd"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.904357 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-run" (OuterVolumeSpecName: "run") pod "b3ea9d97-694e-43f8-8747-64c34bdab8dd" (UID: "b3ea9d97-694e-43f8-8747-64c34bdab8dd"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.904418 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "b3ea9d97-694e-43f8-8747-64c34bdab8dd" (UID: "b3ea9d97-694e-43f8-8747-64c34bdab8dd"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.904449 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "b3ea9d97-694e-43f8-8747-64c34bdab8dd" (UID: "b3ea9d97-694e-43f8-8747-64c34bdab8dd"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.904473 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b3ea9d97-694e-43f8-8747-64c34bdab8dd" (UID: "b3ea9d97-694e-43f8-8747-64c34bdab8dd"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.904495 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "b3ea9d97-694e-43f8-8747-64c34bdab8dd" (UID: "b3ea9d97-694e-43f8-8747-64c34bdab8dd"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.907664 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-scripts" (OuterVolumeSpecName: "scripts") pod "b3ea9d97-694e-43f8-8747-64c34bdab8dd" (UID: "b3ea9d97-694e-43f8-8747-64c34bdab8dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.908014 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3ea9d97-694e-43f8-8747-64c34bdab8dd-kube-api-access-dfbvk" (OuterVolumeSpecName: "kube-api-access-dfbvk") pod "b3ea9d97-694e-43f8-8747-64c34bdab8dd" (UID: "b3ea9d97-694e-43f8-8747-64c34bdab8dd"). InnerVolumeSpecName "kube-api-access-dfbvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.908085 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-lib-cinder\") pod \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.908106 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-sys" (OuterVolumeSpecName: "sys") pod "b3ea9d97-694e-43f8-8747-64c34bdab8dd" (UID: "b3ea9d97-694e-43f8-8747-64c34bdab8dd"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.908136 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-config-data\") pod \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.908161 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "b3ea9d97-694e-43f8-8747-64c34bdab8dd" (UID: "b3ea9d97-694e-43f8-8747-64c34bdab8dd"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.908183 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-combined-ca-bundle\") pod \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\" (UID: \"b3ea9d97-694e-43f8-8747-64c34bdab8dd\") " Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.910033 4706 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.910054 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.910066 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.910078 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-run\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.910090 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfbvk\" (UniqueName: \"kubernetes.io/projected/b3ea9d97-694e-43f8-8747-64c34bdab8dd-kube-api-access-dfbvk\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.910101 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-dev\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.910111 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.910142 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.910152 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.910164 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.910176 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.910188 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b3ea9d97-694e-43f8-8747-64c34bdab8dd-sys\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:10 crc kubenswrapper[4706]: I1206 14:16:10.910431 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b3ea9d97-694e-43f8-8747-64c34bdab8dd" (UID: "b3ea9d97-694e-43f8-8747-64c34bdab8dd"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:10.950230 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3ea9d97-694e-43f8-8747-64c34bdab8dd" (UID: "b3ea9d97-694e-43f8-8747-64c34bdab8dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:10.983464 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-config-data" (OuterVolumeSpecName: "config-data") pod "b3ea9d97-694e-43f8-8747-64c34bdab8dd" (UID: "b3ea9d97-694e-43f8-8747-64c34bdab8dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.016697 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.016724 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.016735 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3ea9d97-694e-43f8-8747-64c34bdab8dd-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.212486 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cindercf71-account-delete-j99bm"] Dec 06 14:16:11 crc kubenswrapper[4706]: W1206 14:16:11.213547 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f400a2f_eb0c_4a3d_a1a3_463b437b33d6.slice/crio-199f115d7807dc634fb81e37dd9393fec3317f549ec144bc38e6bdd88b1d0116 WatchSource:0}: Error finding container 199f115d7807dc634fb81e37dd9393fec3317f549ec144bc38e6bdd88b1d0116: Status 404 returned error can't find the container with id 199f115d7807dc634fb81e37dd9393fec3317f549ec144bc38e6bdd88b1d0116 Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.355582 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cindercf71-account-delete-j99bm" event={"ID":"5f400a2f-eb0c-4a3d-a1a3-463b437b33d6","Type":"ContainerStarted","Data":"dda4728304f8c40611067b964e75b88e7469efe8dde5ec2124e7df8c3daf3639"} Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.355630 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cindercf71-account-delete-j99bm" event={"ID":"5f400a2f-eb0c-4a3d-a1a3-463b437b33d6","Type":"ContainerStarted","Data":"199f115d7807dc634fb81e37dd9393fec3317f549ec144bc38e6bdd88b1d0116"} Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.360003 4706 generic.go:334] "Generic (PLEG): container finished" podID="059b6eae-b9fa-43fd-92a4-f61be1796941" containerID="c85226ca8172c2f9b2711df6efb2637a4b4965da96001b821bbb8e2b00799052" exitCode=0 Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.360071 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"059b6eae-b9fa-43fd-92a4-f61be1796941","Type":"ContainerDied","Data":"c85226ca8172c2f9b2711df6efb2637a4b4965da96001b821bbb8e2b00799052"} Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.362015 4706 generic.go:334] "Generic (PLEG): container finished" podID="5375678a-f779-46ba-bcf4-71932e9ac6c5" containerID="66815d90ff92c4f1709453e09ae9af94254ea4357ac10399dfb261a4a975901c" exitCode=143 Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.362080 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"5375678a-f779-46ba-bcf4-71932e9ac6c5","Type":"ContainerDied","Data":"66815d90ff92c4f1709453e09ae9af94254ea4357ac10399dfb261a4a975901c"} Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.373093 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cindercf71-account-delete-j99bm" podStartSLOduration=1.3730714050000001 podStartE2EDuration="1.373071405s" podCreationTimestamp="2025-12-06 14:16:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:16:11.37026793 +0000 UTC m=+1214.206060797" watchObservedRunningTime="2025-12-06 14:16:11.373071405 +0000 UTC m=+1214.208864282" Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.374810 4706 generic.go:334] "Generic (PLEG): container finished" podID="f10bc859-5282-4d77-8287-823c2f480bc1" containerID="f1875d1e195ab51c83a5ab7dc59f64832d00d51b571cec82c2a8bab8833ac405" exitCode=0 Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.374881 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"f10bc859-5282-4d77-8287-823c2f480bc1","Type":"ContainerDied","Data":"f1875d1e195ab51c83a5ab7dc59f64832d00d51b571cec82c2a8bab8833ac405"} Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.377436 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"b3ea9d97-694e-43f8-8747-64c34bdab8dd","Type":"ContainerDied","Data":"66789396af9bc12e94585b205b3eb9fc307afc81380f9624cdf46a8a86956650"} Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.377483 4706 scope.go:117] "RemoveContainer" containerID="16afbb7105a12e7852686ed389c4ae21ed4ba22f8eeefe72c007f7bc7acafc62" Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.377565 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.410722 4706 scope.go:117] "RemoveContainer" containerID="15687f42b4d952dc619dc822b1a51e51de9aa12b0989f3dfd215c26253896e0f" Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.418159 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.425198 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.518998 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" path="/var/lib/kubelet/pods/b3ea9d97-694e-43f8-8747-64c34bdab8dd/volumes" Dec 06 14:16:11 crc kubenswrapper[4706]: I1206 14:16:11.519680 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cad87657-2720-489c-b6c1-386b6e967c47" path="/var/lib/kubelet/pods/cad87657-2720-489c-b6c1-386b6e967c47/volumes" Dec 06 14:16:12 crc kubenswrapper[4706]: I1206 14:16:12.385759 4706 generic.go:334] "Generic (PLEG): container finished" podID="5f400a2f-eb0c-4a3d-a1a3-463b437b33d6" containerID="dda4728304f8c40611067b964e75b88e7469efe8dde5ec2124e7df8c3daf3639" exitCode=0 Dec 06 14:16:12 crc kubenswrapper[4706]: I1206 14:16:12.385821 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cindercf71-account-delete-j99bm" event={"ID":"5f400a2f-eb0c-4a3d-a1a3-463b437b33d6","Type":"ContainerDied","Data":"dda4728304f8c40611067b964e75b88e7469efe8dde5ec2124e7df8c3daf3639"} Dec 06 14:16:13 crc kubenswrapper[4706]: I1206 14:16:13.748194 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cindercf71-account-delete-j99bm" Dec 06 14:16:13 crc kubenswrapper[4706]: I1206 14:16:13.757795 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdkhd\" (UniqueName: \"kubernetes.io/projected/5f400a2f-eb0c-4a3d-a1a3-463b437b33d6-kube-api-access-hdkhd\") pod \"5f400a2f-eb0c-4a3d-a1a3-463b437b33d6\" (UID: \"5f400a2f-eb0c-4a3d-a1a3-463b437b33d6\") " Dec 06 14:16:13 crc kubenswrapper[4706]: I1206 14:16:13.757910 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f400a2f-eb0c-4a3d-a1a3-463b437b33d6-operator-scripts\") pod \"5f400a2f-eb0c-4a3d-a1a3-463b437b33d6\" (UID: \"5f400a2f-eb0c-4a3d-a1a3-463b437b33d6\") " Dec 06 14:16:13 crc kubenswrapper[4706]: I1206 14:16:13.758848 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f400a2f-eb0c-4a3d-a1a3-463b437b33d6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5f400a2f-eb0c-4a3d-a1a3-463b437b33d6" (UID: "5f400a2f-eb0c-4a3d-a1a3-463b437b33d6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:16:13 crc kubenswrapper[4706]: I1206 14:16:13.767289 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f400a2f-eb0c-4a3d-a1a3-463b437b33d6-kube-api-access-hdkhd" (OuterVolumeSpecName: "kube-api-access-hdkhd") pod "5f400a2f-eb0c-4a3d-a1a3-463b437b33d6" (UID: "5f400a2f-eb0c-4a3d-a1a3-463b437b33d6"). InnerVolumeSpecName "kube-api-access-hdkhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:13 crc kubenswrapper[4706]: I1206 14:16:13.859103 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdkhd\" (UniqueName: \"kubernetes.io/projected/5f400a2f-eb0c-4a3d-a1a3-463b437b33d6-kube-api-access-hdkhd\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:13 crc kubenswrapper[4706]: I1206 14:16:13.859137 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f400a2f-eb0c-4a3d-a1a3-463b437b33d6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.310653 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.358115 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367212 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-dev\") pod \"f10bc859-5282-4d77-8287-823c2f480bc1\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367269 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc55b\" (UniqueName: \"kubernetes.io/projected/059b6eae-b9fa-43fd-92a4-f61be1796941-kube-api-access-nc55b\") pod \"059b6eae-b9fa-43fd-92a4-f61be1796941\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367354 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-lib-modules\") pod \"f10bc859-5282-4d77-8287-823c2f480bc1\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367357 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-dev" (OuterVolumeSpecName: "dev") pod "f10bc859-5282-4d77-8287-823c2f480bc1" (UID: "f10bc859-5282-4d77-8287-823c2f480bc1"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367376 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-locks-brick\") pod \"f10bc859-5282-4d77-8287-823c2f480bc1\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367432 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "f10bc859-5282-4d77-8287-823c2f480bc1" (UID: "f10bc859-5282-4d77-8287-823c2f480bc1"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367439 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-run\") pod \"f10bc859-5282-4d77-8287-823c2f480bc1\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367474 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-config-data-custom\") pod \"f10bc859-5282-4d77-8287-823c2f480bc1\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367507 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/059b6eae-b9fa-43fd-92a4-f61be1796941-etc-machine-id\") pod \"059b6eae-b9fa-43fd-92a4-f61be1796941\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367502 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "f10bc859-5282-4d77-8287-823c2f480bc1" (UID: "f10bc859-5282-4d77-8287-823c2f480bc1"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367529 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-nvme\") pod \"f10bc859-5282-4d77-8287-823c2f480bc1\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367562 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-config-data\") pod \"f10bc859-5282-4d77-8287-823c2f480bc1\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367577 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "f10bc859-5282-4d77-8287-823c2f480bc1" (UID: "f10bc859-5282-4d77-8287-823c2f480bc1"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367599 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-config-data-custom\") pod \"059b6eae-b9fa-43fd-92a4-f61be1796941\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367615 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/059b6eae-b9fa-43fd-92a4-f61be1796941-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "059b6eae-b9fa-43fd-92a4-f61be1796941" (UID: "059b6eae-b9fa-43fd-92a4-f61be1796941"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367632 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-config-data\") pod \"059b6eae-b9fa-43fd-92a4-f61be1796941\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367661 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-iscsi\") pod \"f10bc859-5282-4d77-8287-823c2f480bc1\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367665 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-run" (OuterVolumeSpecName: "run") pod "f10bc859-5282-4d77-8287-823c2f480bc1" (UID: "f10bc859-5282-4d77-8287-823c2f480bc1"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367697 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-combined-ca-bundle\") pod \"059b6eae-b9fa-43fd-92a4-f61be1796941\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367725 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-lib-cinder\") pod \"f10bc859-5282-4d77-8287-823c2f480bc1\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367741 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "f10bc859-5282-4d77-8287-823c2f480bc1" (UID: "f10bc859-5282-4d77-8287-823c2f480bc1"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367747 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqdz5\" (UniqueName: \"kubernetes.io/projected/f10bc859-5282-4d77-8287-823c2f480bc1-kube-api-access-xqdz5\") pod \"f10bc859-5282-4d77-8287-823c2f480bc1\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367799 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-machine-id\") pod \"f10bc859-5282-4d77-8287-823c2f480bc1\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367886 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-scripts\") pod \"059b6eae-b9fa-43fd-92a4-f61be1796941\" (UID: \"059b6eae-b9fa-43fd-92a4-f61be1796941\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367931 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-combined-ca-bundle\") pod \"f10bc859-5282-4d77-8287-823c2f480bc1\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.367971 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-sys\") pod \"f10bc859-5282-4d77-8287-823c2f480bc1\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.368029 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-locks-cinder\") pod \"f10bc859-5282-4d77-8287-823c2f480bc1\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.368060 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-scripts\") pod \"f10bc859-5282-4d77-8287-823c2f480bc1\" (UID: \"f10bc859-5282-4d77-8287-823c2f480bc1\") " Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.368515 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.368540 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-dev\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.368556 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.368573 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.368588 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-run\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.368604 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/059b6eae-b9fa-43fd-92a4-f61be1796941-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.368621 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.369487 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-sys" (OuterVolumeSpecName: "sys") pod "f10bc859-5282-4d77-8287-823c2f480bc1" (UID: "f10bc859-5282-4d77-8287-823c2f480bc1"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.369526 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "f10bc859-5282-4d77-8287-823c2f480bc1" (UID: "f10bc859-5282-4d77-8287-823c2f480bc1"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.369559 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "f10bc859-5282-4d77-8287-823c2f480bc1" (UID: "f10bc859-5282-4d77-8287-823c2f480bc1"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.369586 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f10bc859-5282-4d77-8287-823c2f480bc1" (UID: "f10bc859-5282-4d77-8287-823c2f480bc1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.372896 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f10bc859-5282-4d77-8287-823c2f480bc1" (UID: "f10bc859-5282-4d77-8287-823c2f480bc1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.377537 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "059b6eae-b9fa-43fd-92a4-f61be1796941" (UID: "059b6eae-b9fa-43fd-92a4-f61be1796941"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.377563 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f10bc859-5282-4d77-8287-823c2f480bc1-kube-api-access-xqdz5" (OuterVolumeSpecName: "kube-api-access-xqdz5") pod "f10bc859-5282-4d77-8287-823c2f480bc1" (UID: "f10bc859-5282-4d77-8287-823c2f480bc1"). InnerVolumeSpecName "kube-api-access-xqdz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.377616 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/059b6eae-b9fa-43fd-92a4-f61be1796941-kube-api-access-nc55b" (OuterVolumeSpecName: "kube-api-access-nc55b") pod "059b6eae-b9fa-43fd-92a4-f61be1796941" (UID: "059b6eae-b9fa-43fd-92a4-f61be1796941"). InnerVolumeSpecName "kube-api-access-nc55b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.377944 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-scripts" (OuterVolumeSpecName: "scripts") pod "f10bc859-5282-4d77-8287-823c2f480bc1" (UID: "f10bc859-5282-4d77-8287-823c2f480bc1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.378720 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-scripts" (OuterVolumeSpecName: "scripts") pod "059b6eae-b9fa-43fd-92a4-f61be1796941" (UID: "059b6eae-b9fa-43fd-92a4-f61be1796941"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.407032 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cindercf71-account-delete-j99bm" event={"ID":"5f400a2f-eb0c-4a3d-a1a3-463b437b33d6","Type":"ContainerDied","Data":"199f115d7807dc634fb81e37dd9393fec3317f549ec144bc38e6bdd88b1d0116"} Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.407075 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cindercf71-account-delete-j99bm" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.407102 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="199f115d7807dc634fb81e37dd9393fec3317f549ec144bc38e6bdd88b1d0116" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.408921 4706 generic.go:334] "Generic (PLEG): container finished" podID="f10bc859-5282-4d77-8287-823c2f480bc1" containerID="103ec7c6f4792616852ef5c8055c1871f9d15d6bdf195f49a46f5ce642f0d340" exitCode=0 Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.408991 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"f10bc859-5282-4d77-8287-823c2f480bc1","Type":"ContainerDied","Data":"103ec7c6f4792616852ef5c8055c1871f9d15d6bdf195f49a46f5ce642f0d340"} Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.409014 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.409029 4706 scope.go:117] "RemoveContainer" containerID="f1875d1e195ab51c83a5ab7dc59f64832d00d51b571cec82c2a8bab8833ac405" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.409015 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"f10bc859-5282-4d77-8287-823c2f480bc1","Type":"ContainerDied","Data":"6abe933ab9e15b5cd35821769c8bc23a562ec307dbbd0c6f09b79f66e04ff01a"} Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.410905 4706 generic.go:334] "Generic (PLEG): container finished" podID="059b6eae-b9fa-43fd-92a4-f61be1796941" containerID="b24aaf8d7884a0230176a110498ef30b0704f3d3b9162a0773117e03b098fd9c" exitCode=0 Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.410929 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"059b6eae-b9fa-43fd-92a4-f61be1796941","Type":"ContainerDied","Data":"b24aaf8d7884a0230176a110498ef30b0704f3d3b9162a0773117e03b098fd9c"} Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.410948 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"059b6eae-b9fa-43fd-92a4-f61be1796941","Type":"ContainerDied","Data":"a3196368460110a6ca17e240b421e254d95e70eaf8ad91b3ed4055e738651ece"} Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.410970 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.411827 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "059b6eae-b9fa-43fd-92a4-f61be1796941" (UID: "059b6eae-b9fa-43fd-92a4-f61be1796941"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.424920 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f10bc859-5282-4d77-8287-823c2f480bc1" (UID: "f10bc859-5282-4d77-8287-823c2f480bc1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.441518 4706 scope.go:117] "RemoveContainer" containerID="103ec7c6f4792616852ef5c8055c1871f9d15d6bdf195f49a46f5ce642f0d340" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.460146 4706 scope.go:117] "RemoveContainer" containerID="f1875d1e195ab51c83a5ab7dc59f64832d00d51b571cec82c2a8bab8833ac405" Dec 06 14:16:14 crc kubenswrapper[4706]: E1206 14:16:14.460552 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1875d1e195ab51c83a5ab7dc59f64832d00d51b571cec82c2a8bab8833ac405\": container with ID starting with f1875d1e195ab51c83a5ab7dc59f64832d00d51b571cec82c2a8bab8833ac405 not found: ID does not exist" containerID="f1875d1e195ab51c83a5ab7dc59f64832d00d51b571cec82c2a8bab8833ac405" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.460595 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1875d1e195ab51c83a5ab7dc59f64832d00d51b571cec82c2a8bab8833ac405"} err="failed to get container status \"f1875d1e195ab51c83a5ab7dc59f64832d00d51b571cec82c2a8bab8833ac405\": rpc error: code = NotFound desc = could not find container \"f1875d1e195ab51c83a5ab7dc59f64832d00d51b571cec82c2a8bab8833ac405\": container with ID starting with f1875d1e195ab51c83a5ab7dc59f64832d00d51b571cec82c2a8bab8833ac405 not found: ID does not exist" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.460623 4706 scope.go:117] "RemoveContainer" containerID="103ec7c6f4792616852ef5c8055c1871f9d15d6bdf195f49a46f5ce642f0d340" Dec 06 14:16:14 crc kubenswrapper[4706]: E1206 14:16:14.460900 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"103ec7c6f4792616852ef5c8055c1871f9d15d6bdf195f49a46f5ce642f0d340\": container with ID starting with 103ec7c6f4792616852ef5c8055c1871f9d15d6bdf195f49a46f5ce642f0d340 not found: ID does not exist" containerID="103ec7c6f4792616852ef5c8055c1871f9d15d6bdf195f49a46f5ce642f0d340" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.460935 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"103ec7c6f4792616852ef5c8055c1871f9d15d6bdf195f49a46f5ce642f0d340"} err="failed to get container status \"103ec7c6f4792616852ef5c8055c1871f9d15d6bdf195f49a46f5ce642f0d340\": rpc error: code = NotFound desc = could not find container \"103ec7c6f4792616852ef5c8055c1871f9d15d6bdf195f49a46f5ce642f0d340\": container with ID starting with 103ec7c6f4792616852ef5c8055c1871f9d15d6bdf195f49a46f5ce642f0d340 not found: ID does not exist" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.460958 4706 scope.go:117] "RemoveContainer" containerID="c85226ca8172c2f9b2711df6efb2637a4b4965da96001b821bbb8e2b00799052" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.461250 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-config-data" (OuterVolumeSpecName: "config-data") pod "059b6eae-b9fa-43fd-92a4-f61be1796941" (UID: "059b6eae-b9fa-43fd-92a4-f61be1796941"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.468008 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-config-data" (OuterVolumeSpecName: "config-data") pod "f10bc859-5282-4d77-8287-823c2f480bc1" (UID: "f10bc859-5282-4d77-8287-823c2f480bc1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.469100 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.469122 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.469134 4706 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.469144 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqdz5\" (UniqueName: \"kubernetes.io/projected/f10bc859-5282-4d77-8287-823c2f480bc1-kube-api-access-xqdz5\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.469154 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.469164 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.469173 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.469181 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-sys\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.469190 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f10bc859-5282-4d77-8287-823c2f480bc1-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.469198 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.469205 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc55b\" (UniqueName: \"kubernetes.io/projected/059b6eae-b9fa-43fd-92a4-f61be1796941-kube-api-access-nc55b\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.469213 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.469222 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10bc859-5282-4d77-8287-823c2f480bc1-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.469231 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/059b6eae-b9fa-43fd-92a4-f61be1796941-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.478205 4706 scope.go:117] "RemoveContainer" containerID="b24aaf8d7884a0230176a110498ef30b0704f3d3b9162a0773117e03b098fd9c" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.493863 4706 scope.go:117] "RemoveContainer" containerID="c85226ca8172c2f9b2711df6efb2637a4b4965da96001b821bbb8e2b00799052" Dec 06 14:16:14 crc kubenswrapper[4706]: E1206 14:16:14.494250 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c85226ca8172c2f9b2711df6efb2637a4b4965da96001b821bbb8e2b00799052\": container with ID starting with c85226ca8172c2f9b2711df6efb2637a4b4965da96001b821bbb8e2b00799052 not found: ID does not exist" containerID="c85226ca8172c2f9b2711df6efb2637a4b4965da96001b821bbb8e2b00799052" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.494292 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c85226ca8172c2f9b2711df6efb2637a4b4965da96001b821bbb8e2b00799052"} err="failed to get container status \"c85226ca8172c2f9b2711df6efb2637a4b4965da96001b821bbb8e2b00799052\": rpc error: code = NotFound desc = could not find container \"c85226ca8172c2f9b2711df6efb2637a4b4965da96001b821bbb8e2b00799052\": container with ID starting with c85226ca8172c2f9b2711df6efb2637a4b4965da96001b821bbb8e2b00799052 not found: ID does not exist" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.494339 4706 scope.go:117] "RemoveContainer" containerID="b24aaf8d7884a0230176a110498ef30b0704f3d3b9162a0773117e03b098fd9c" Dec 06 14:16:14 crc kubenswrapper[4706]: E1206 14:16:14.494684 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b24aaf8d7884a0230176a110498ef30b0704f3d3b9162a0773117e03b098fd9c\": container with ID starting with b24aaf8d7884a0230176a110498ef30b0704f3d3b9162a0773117e03b098fd9c not found: ID does not exist" containerID="b24aaf8d7884a0230176a110498ef30b0704f3d3b9162a0773117e03b098fd9c" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.494712 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b24aaf8d7884a0230176a110498ef30b0704f3d3b9162a0773117e03b098fd9c"} err="failed to get container status \"b24aaf8d7884a0230176a110498ef30b0704f3d3b9162a0773117e03b098fd9c\": rpc error: code = NotFound desc = could not find container \"b24aaf8d7884a0230176a110498ef30b0704f3d3b9162a0773117e03b098fd9c\": container with ID starting with b24aaf8d7884a0230176a110498ef30b0704f3d3b9162a0773117e03b098fd9c not found: ID does not exist" Dec 06 14:16:14 crc kubenswrapper[4706]: E1206 14:16:14.550695 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f400a2f_eb0c_4a3d_a1a3_463b437b33d6.slice\": RecentStats: unable to find data in memory cache]" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.580518 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.580594 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.580646 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.581410 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"953455e6657a15cff411732f38c90183866bd015ab24e23297c405658c8efab6"} pod="openshift-machine-config-operator/machine-config-daemon-67v56" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.581490 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" containerID="cri-o://953455e6657a15cff411732f38c90183866bd015ab24e23297c405658c8efab6" gracePeriod=600 Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.749163 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.760996 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.792673 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.799140 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 06 14:16:14 crc kubenswrapper[4706]: I1206 14:16:14.902800 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-0" podUID="5375678a-f779-46ba-bcf4-71932e9ac6c5" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.108:8776/healthcheck\": read tcp 10.217.0.2:57982->10.217.0.108:8776: read: connection reset by peer" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.336200 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.382994 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-config-data\") pod \"5375678a-f779-46ba-bcf4-71932e9ac6c5\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.383061 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-combined-ca-bundle\") pod \"5375678a-f779-46ba-bcf4-71932e9ac6c5\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.383184 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5375678a-f779-46ba-bcf4-71932e9ac6c5-etc-machine-id\") pod \"5375678a-f779-46ba-bcf4-71932e9ac6c5\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.383232 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkjrk\" (UniqueName: \"kubernetes.io/projected/5375678a-f779-46ba-bcf4-71932e9ac6c5-kube-api-access-dkjrk\") pod \"5375678a-f779-46ba-bcf4-71932e9ac6c5\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.383363 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-config-data-custom\") pod \"5375678a-f779-46ba-bcf4-71932e9ac6c5\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.383395 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5375678a-f779-46ba-bcf4-71932e9ac6c5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5375678a-f779-46ba-bcf4-71932e9ac6c5" (UID: "5375678a-f779-46ba-bcf4-71932e9ac6c5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.383425 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5375678a-f779-46ba-bcf4-71932e9ac6c5-logs\") pod \"5375678a-f779-46ba-bcf4-71932e9ac6c5\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.383534 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-public-tls-certs\") pod \"5375678a-f779-46ba-bcf4-71932e9ac6c5\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.383577 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-scripts\") pod \"5375678a-f779-46ba-bcf4-71932e9ac6c5\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.383622 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-internal-tls-certs\") pod \"5375678a-f779-46ba-bcf4-71932e9ac6c5\" (UID: \"5375678a-f779-46ba-bcf4-71932e9ac6c5\") " Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.384269 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5375678a-f779-46ba-bcf4-71932e9ac6c5-logs" (OuterVolumeSpecName: "logs") pod "5375678a-f779-46ba-bcf4-71932e9ac6c5" (UID: "5375678a-f779-46ba-bcf4-71932e9ac6c5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.384604 4706 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5375678a-f779-46ba-bcf4-71932e9ac6c5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.384634 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5375678a-f779-46ba-bcf4-71932e9ac6c5-logs\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.390180 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-scripts" (OuterVolumeSpecName: "scripts") pod "5375678a-f779-46ba-bcf4-71932e9ac6c5" (UID: "5375678a-f779-46ba-bcf4-71932e9ac6c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.390398 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5375678a-f779-46ba-bcf4-71932e9ac6c5-kube-api-access-dkjrk" (OuterVolumeSpecName: "kube-api-access-dkjrk") pod "5375678a-f779-46ba-bcf4-71932e9ac6c5" (UID: "5375678a-f779-46ba-bcf4-71932e9ac6c5"). InnerVolumeSpecName "kube-api-access-dkjrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.390671 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5375678a-f779-46ba-bcf4-71932e9ac6c5" (UID: "5375678a-f779-46ba-bcf4-71932e9ac6c5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.407606 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5375678a-f779-46ba-bcf4-71932e9ac6c5" (UID: "5375678a-f779-46ba-bcf4-71932e9ac6c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.431914 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5375678a-f779-46ba-bcf4-71932e9ac6c5" (UID: "5375678a-f779-46ba-bcf4-71932e9ac6c5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.443827 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-config-data" (OuterVolumeSpecName: "config-data") pod "5375678a-f779-46ba-bcf4-71932e9ac6c5" (UID: "5375678a-f779-46ba-bcf4-71932e9ac6c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.443896 4706 generic.go:334] "Generic (PLEG): container finished" podID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerID="953455e6657a15cff411732f38c90183866bd015ab24e23297c405658c8efab6" exitCode=0 Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.443979 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerDied","Data":"953455e6657a15cff411732f38c90183866bd015ab24e23297c405658c8efab6"} Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.444018 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerStarted","Data":"541fe7945667e62148e85f3104c224b6e7d1a0ff6a7b9516e33ef2027678744f"} Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.444083 4706 scope.go:117] "RemoveContainer" containerID="fc4d148c623467502d8b194b1ce686e138f01f9dcba1a1b63282c7d88ff885f4" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.449687 4706 generic.go:334] "Generic (PLEG): container finished" podID="5375678a-f779-46ba-bcf4-71932e9ac6c5" containerID="34fb37f23daecfa827f3d223a2d3f2a11c676aca2295b0bcba9329c001b0dbe3" exitCode=0 Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.449751 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"5375678a-f779-46ba-bcf4-71932e9ac6c5","Type":"ContainerDied","Data":"34fb37f23daecfa827f3d223a2d3f2a11c676aca2295b0bcba9329c001b0dbe3"} Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.449777 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"5375678a-f779-46ba-bcf4-71932e9ac6c5","Type":"ContainerDied","Data":"8c5bfcbedb802b8f2a803098e5de887d06b1903a1266ca736ab852c4931fcf0b"} Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.449795 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.461614 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5375678a-f779-46ba-bcf4-71932e9ac6c5" (UID: "5375678a-f779-46ba-bcf4-71932e9ac6c5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.462563 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-8kvhz"] Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.482412 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-8kvhz"] Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.486985 4706 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.487007 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.487543 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.487573 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkjrk\" (UniqueName: \"kubernetes.io/projected/5375678a-f779-46ba-bcf4-71932e9ac6c5-kube-api-access-dkjrk\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.487588 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.487600 4706 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.487613 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5375678a-f779-46ba-bcf4-71932e9ac6c5-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.490652 4706 scope.go:117] "RemoveContainer" containerID="34fb37f23daecfa827f3d223a2d3f2a11c676aca2295b0bcba9329c001b0dbe3" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.494715 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz"] Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.503141 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cindercf71-account-delete-j99bm"] Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.509235 4706 scope.go:117] "RemoveContainer" containerID="66815d90ff92c4f1709453e09ae9af94254ea4357ac10399dfb261a4a975901c" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.519588 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="059b6eae-b9fa-43fd-92a4-f61be1796941" path="/var/lib/kubelet/pods/059b6eae-b9fa-43fd-92a4-f61be1796941/volumes" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.520527 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f10bc859-5282-4d77-8287-823c2f480bc1" path="/var/lib/kubelet/pods/f10bc859-5282-4d77-8287-823c2f480bc1/volumes" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.521257 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99" path="/var/lib/kubelet/pods/f7f0f60f-9f01-4d0e-aaf7-4908d2d0fb99/volumes" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.522883 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-cf71-account-create-update-svhfz"] Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.522921 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cindercf71-account-delete-j99bm"] Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.526125 4706 scope.go:117] "RemoveContainer" containerID="34fb37f23daecfa827f3d223a2d3f2a11c676aca2295b0bcba9329c001b0dbe3" Dec 06 14:16:15 crc kubenswrapper[4706]: E1206 14:16:15.526728 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34fb37f23daecfa827f3d223a2d3f2a11c676aca2295b0bcba9329c001b0dbe3\": container with ID starting with 34fb37f23daecfa827f3d223a2d3f2a11c676aca2295b0bcba9329c001b0dbe3 not found: ID does not exist" containerID="34fb37f23daecfa827f3d223a2d3f2a11c676aca2295b0bcba9329c001b0dbe3" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.526772 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34fb37f23daecfa827f3d223a2d3f2a11c676aca2295b0bcba9329c001b0dbe3"} err="failed to get container status \"34fb37f23daecfa827f3d223a2d3f2a11c676aca2295b0bcba9329c001b0dbe3\": rpc error: code = NotFound desc = could not find container \"34fb37f23daecfa827f3d223a2d3f2a11c676aca2295b0bcba9329c001b0dbe3\": container with ID starting with 34fb37f23daecfa827f3d223a2d3f2a11c676aca2295b0bcba9329c001b0dbe3 not found: ID does not exist" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.526802 4706 scope.go:117] "RemoveContainer" containerID="66815d90ff92c4f1709453e09ae9af94254ea4357ac10399dfb261a4a975901c" Dec 06 14:16:15 crc kubenswrapper[4706]: E1206 14:16:15.527122 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66815d90ff92c4f1709453e09ae9af94254ea4357ac10399dfb261a4a975901c\": container with ID starting with 66815d90ff92c4f1709453e09ae9af94254ea4357ac10399dfb261a4a975901c not found: ID does not exist" containerID="66815d90ff92c4f1709453e09ae9af94254ea4357ac10399dfb261a4a975901c" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.527161 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66815d90ff92c4f1709453e09ae9af94254ea4357ac10399dfb261a4a975901c"} err="failed to get container status \"66815d90ff92c4f1709453e09ae9af94254ea4357ac10399dfb261a4a975901c\": rpc error: code = NotFound desc = could not find container \"66815d90ff92c4f1709453e09ae9af94254ea4357ac10399dfb261a4a975901c\": container with ID starting with 66815d90ff92c4f1709453e09ae9af94254ea4357ac10399dfb261a4a975901c not found: ID does not exist" Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.783445 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:16:15 crc kubenswrapper[4706]: I1206 14:16:15.792765 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.909549 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-fzrvx"] Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.918665 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-fzrvx"] Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.923103 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-ls6wk"] Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.928336 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-ls6wk"] Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.932333 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh"] Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.932568 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" podUID="4ddad38d-0652-48ab-a70e-7eca376ba89d" containerName="keystone-api" containerID="cri-o://22412e6cb2fd24e10cf3f94722611789ed23e1b4208f9601d256eca63c15b73c" gracePeriod=30 Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.970868 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystoneec25-account-delete-47t8b"] Dec 06 14:16:16 crc kubenswrapper[4706]: E1206 14:16:16.971156 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" containerName="probe" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971174 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" containerName="probe" Dec 06 14:16:16 crc kubenswrapper[4706]: E1206 14:16:16.971184 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="059b6eae-b9fa-43fd-92a4-f61be1796941" containerName="probe" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971191 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="059b6eae-b9fa-43fd-92a4-f61be1796941" containerName="probe" Dec 06 14:16:16 crc kubenswrapper[4706]: E1206 14:16:16.971206 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5375678a-f779-46ba-bcf4-71932e9ac6c5" containerName="cinder-api" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971212 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5375678a-f779-46ba-bcf4-71932e9ac6c5" containerName="cinder-api" Dec 06 14:16:16 crc kubenswrapper[4706]: E1206 14:16:16.971223 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" containerName="probe" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971228 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" containerName="probe" Dec 06 14:16:16 crc kubenswrapper[4706]: E1206 14:16:16.971237 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f400a2f-eb0c-4a3d-a1a3-463b437b33d6" containerName="mariadb-account-delete" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971244 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f400a2f-eb0c-4a3d-a1a3-463b437b33d6" containerName="mariadb-account-delete" Dec 06 14:16:16 crc kubenswrapper[4706]: E1206 14:16:16.971254 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5375678a-f779-46ba-bcf4-71932e9ac6c5" containerName="cinder-api-log" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971259 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5375678a-f779-46ba-bcf4-71932e9ac6c5" containerName="cinder-api-log" Dec 06 14:16:16 crc kubenswrapper[4706]: E1206 14:16:16.971267 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" containerName="cinder-volume" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971272 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" containerName="cinder-volume" Dec 06 14:16:16 crc kubenswrapper[4706]: E1206 14:16:16.971280 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" containerName="cinder-volume" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971286 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" containerName="cinder-volume" Dec 06 14:16:16 crc kubenswrapper[4706]: E1206 14:16:16.971296 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f10bc859-5282-4d77-8287-823c2f480bc1" containerName="probe" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971302 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f10bc859-5282-4d77-8287-823c2f480bc1" containerName="probe" Dec 06 14:16:16 crc kubenswrapper[4706]: E1206 14:16:16.971327 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="059b6eae-b9fa-43fd-92a4-f61be1796941" containerName="cinder-scheduler" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971334 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="059b6eae-b9fa-43fd-92a4-f61be1796941" containerName="cinder-scheduler" Dec 06 14:16:16 crc kubenswrapper[4706]: E1206 14:16:16.971344 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f10bc859-5282-4d77-8287-823c2f480bc1" containerName="cinder-backup" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971350 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f10bc859-5282-4d77-8287-823c2f480bc1" containerName="cinder-backup" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971479 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f400a2f-eb0c-4a3d-a1a3-463b437b33d6" containerName="mariadb-account-delete" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971497 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5375678a-f779-46ba-bcf4-71932e9ac6c5" containerName="cinder-api" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971504 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="059b6eae-b9fa-43fd-92a4-f61be1796941" containerName="cinder-scheduler" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971513 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" containerName="cinder-volume" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971519 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5375678a-f779-46ba-bcf4-71932e9ac6c5" containerName="cinder-api-log" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971528 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="059b6eae-b9fa-43fd-92a4-f61be1796941" containerName="probe" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971537 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f10bc859-5282-4d77-8287-823c2f480bc1" containerName="probe" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971545 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" containerName="probe" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971555 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f10bc859-5282-4d77-8287-823c2f480bc1" containerName="cinder-backup" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.971563 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" containerName="probe" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.972020 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" Dec 06 14:16:16 crc kubenswrapper[4706]: I1206 14:16:16.988503 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystoneec25-account-delete-47t8b"] Dec 06 14:16:17 crc kubenswrapper[4706]: I1206 14:16:17.011556 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5dce085-c674-49c0-883c-d1d788c1b70d-operator-scripts\") pod \"keystoneec25-account-delete-47t8b\" (UID: \"d5dce085-c674-49c0-883c-d1d788c1b70d\") " pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" Dec 06 14:16:17 crc kubenswrapper[4706]: I1206 14:16:17.011729 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d6km\" (UniqueName: \"kubernetes.io/projected/d5dce085-c674-49c0-883c-d1d788c1b70d-kube-api-access-5d6km\") pod \"keystoneec25-account-delete-47t8b\" (UID: \"d5dce085-c674-49c0-883c-d1d788c1b70d\") " pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" Dec 06 14:16:17 crc kubenswrapper[4706]: I1206 14:16:17.113116 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5dce085-c674-49c0-883c-d1d788c1b70d-operator-scripts\") pod \"keystoneec25-account-delete-47t8b\" (UID: \"d5dce085-c674-49c0-883c-d1d788c1b70d\") " pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" Dec 06 14:16:17 crc kubenswrapper[4706]: I1206 14:16:17.113235 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d6km\" (UniqueName: \"kubernetes.io/projected/d5dce085-c674-49c0-883c-d1d788c1b70d-kube-api-access-5d6km\") pod \"keystoneec25-account-delete-47t8b\" (UID: \"d5dce085-c674-49c0-883c-d1d788c1b70d\") " pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" Dec 06 14:16:17 crc kubenswrapper[4706]: I1206 14:16:17.113919 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5dce085-c674-49c0-883c-d1d788c1b70d-operator-scripts\") pod \"keystoneec25-account-delete-47t8b\" (UID: \"d5dce085-c674-49c0-883c-d1d788c1b70d\") " pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" Dec 06 14:16:17 crc kubenswrapper[4706]: I1206 14:16:17.133277 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d6km\" (UniqueName: \"kubernetes.io/projected/d5dce085-c674-49c0-883c-d1d788c1b70d-kube-api-access-5d6km\") pod \"keystoneec25-account-delete-47t8b\" (UID: \"d5dce085-c674-49c0-883c-d1d788c1b70d\") " pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" Dec 06 14:16:17 crc kubenswrapper[4706]: I1206 14:16:17.299799 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" Dec 06 14:16:17 crc kubenswrapper[4706]: I1206 14:16:17.524296 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="088ef3b1-37c7-4bcd-b508-847327dc1d38" path="/var/lib/kubelet/pods/088ef3b1-37c7-4bcd-b508-847327dc1d38/volumes" Dec 06 14:16:17 crc kubenswrapper[4706]: I1206 14:16:17.525292 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5375678a-f779-46ba-bcf4-71932e9ac6c5" path="/var/lib/kubelet/pods/5375678a-f779-46ba-bcf4-71932e9ac6c5/volumes" Dec 06 14:16:17 crc kubenswrapper[4706]: I1206 14:16:17.526515 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f400a2f-eb0c-4a3d-a1a3-463b437b33d6" path="/var/lib/kubelet/pods/5f400a2f-eb0c-4a3d-a1a3-463b437b33d6/volumes" Dec 06 14:16:17 crc kubenswrapper[4706]: I1206 14:16:17.527965 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7af82ce5-6c72-4654-b677-f517f8885495" path="/var/lib/kubelet/pods/7af82ce5-6c72-4654-b677-f517f8885495/volumes" Dec 06 14:16:17 crc kubenswrapper[4706]: I1206 14:16:17.528843 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1a69216-9494-48d4-828a-2a0069314d37" path="/var/lib/kubelet/pods/c1a69216-9494-48d4-828a-2a0069314d37/volumes" Dec 06 14:16:17 crc kubenswrapper[4706]: I1206 14:16:17.810990 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 06 14:16:17 crc kubenswrapper[4706]: I1206 14:16:17.824102 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 06 14:16:17 crc kubenswrapper[4706]: I1206 14:16:17.831248 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 06 14:16:17 crc kubenswrapper[4706]: I1206 14:16:17.837445 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystoneec25-account-delete-47t8b"] Dec 06 14:16:17 crc kubenswrapper[4706]: I1206 14:16:17.977051 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/openstack-galera-2" podUID="41b63157-bdec-47c2-a4b0-228f89541a2a" containerName="galera" containerID="cri-o://56494e751042f56606058af3e4fca29a89a08904e452133514ddd5ed3269298a" gracePeriod=30 Dec 06 14:16:18 crc kubenswrapper[4706]: I1206 14:16:18.478845 4706 generic.go:334] "Generic (PLEG): container finished" podID="d5dce085-c674-49c0-883c-d1d788c1b70d" containerID="3bed485b02f207bacdc08706926eb5cc6e3405abfc5bd01d95acd6b174d2c678" exitCode=1 Dec 06 14:16:18 crc kubenswrapper[4706]: I1206 14:16:18.478956 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" event={"ID":"d5dce085-c674-49c0-883c-d1d788c1b70d","Type":"ContainerDied","Data":"3bed485b02f207bacdc08706926eb5cc6e3405abfc5bd01d95acd6b174d2c678"} Dec 06 14:16:18 crc kubenswrapper[4706]: I1206 14:16:18.479243 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" event={"ID":"d5dce085-c674-49c0-883c-d1d788c1b70d","Type":"ContainerStarted","Data":"ccdc38bcab851c50cdfc7003f71b98c26fb5f2faa16f930d5340504ba2c7a6c2"} Dec 06 14:16:18 crc kubenswrapper[4706]: I1206 14:16:18.479375 4706 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" secret="" err="secret \"galera-openstack-dockercfg-4j5n7\" not found" Dec 06 14:16:18 crc kubenswrapper[4706]: I1206 14:16:18.479422 4706 scope.go:117] "RemoveContainer" containerID="3bed485b02f207bacdc08706926eb5cc6e3405abfc5bd01d95acd6b174d2c678" Dec 06 14:16:18 crc kubenswrapper[4706]: I1206 14:16:18.533046 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 06 14:16:18 crc kubenswrapper[4706]: I1206 14:16:18.533372 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/memcached-0" podUID="76a993fd-482d-4efd-938a-2014dda4aca1" containerName="memcached" containerID="cri-o://9f53871a8284673c6d9414d0374c0bf6afe88dd86fc2f4264ab391d6f28557f2" gracePeriod=30 Dec 06 14:16:18 crc kubenswrapper[4706]: E1206 14:16:18.633472 4706 configmap.go:193] Couldn't get configMap cinder-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Dec 06 14:16:18 crc kubenswrapper[4706]: E1206 14:16:18.633763 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d5dce085-c674-49c0-883c-d1d788c1b70d-operator-scripts podName:d5dce085-c674-49c0-883c-d1d788c1b70d nodeName:}" failed. No retries permitted until 2025-12-06 14:16:19.133745446 +0000 UTC m=+1221.969538313 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/d5dce085-c674-49c0-883c-d1d788c1b70d-operator-scripts") pod "keystoneec25-account-delete-47t8b" (UID: "d5dce085-c674-49c0-883c-d1d788c1b70d") : configmap "openstack-scripts" not found Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.052353 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 06 14:16:19 crc kubenswrapper[4706]: E1206 14:16:19.140112 4706 configmap.go:193] Couldn't get configMap cinder-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Dec 06 14:16:19 crc kubenswrapper[4706]: E1206 14:16:19.140172 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d5dce085-c674-49c0-883c-d1d788c1b70d-operator-scripts podName:d5dce085-c674-49c0-883c-d1d788c1b70d nodeName:}" failed. No retries permitted until 2025-12-06 14:16:20.140158576 +0000 UTC m=+1222.975951433 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/d5dce085-c674-49c0-883c-d1d788c1b70d-operator-scripts") pod "keystoneec25-account-delete-47t8b" (UID: "d5dce085-c674-49c0-883c-d1d788c1b70d") : configmap "openstack-scripts" not found Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.254536 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.433869 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.444098 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-operator-scripts\") pod \"41b63157-bdec-47c2-a4b0-228f89541a2a\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.444137 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbgw2\" (UniqueName: \"kubernetes.io/projected/41b63157-bdec-47c2-a4b0-228f89541a2a-kube-api-access-rbgw2\") pod \"41b63157-bdec-47c2-a4b0-228f89541a2a\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.444194 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-kolla-config\") pod \"41b63157-bdec-47c2-a4b0-228f89541a2a\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.444225 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-config-data-default\") pod \"41b63157-bdec-47c2-a4b0-228f89541a2a\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.444283 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/41b63157-bdec-47c2-a4b0-228f89541a2a-config-data-generated\") pod \"41b63157-bdec-47c2-a4b0-228f89541a2a\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.444345 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"41b63157-bdec-47c2-a4b0-228f89541a2a\" (UID: \"41b63157-bdec-47c2-a4b0-228f89541a2a\") " Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.444727 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41b63157-bdec-47c2-a4b0-228f89541a2a-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "41b63157-bdec-47c2-a4b0-228f89541a2a" (UID: "41b63157-bdec-47c2-a4b0-228f89541a2a"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.444844 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "41b63157-bdec-47c2-a4b0-228f89541a2a" (UID: "41b63157-bdec-47c2-a4b0-228f89541a2a"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.444924 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "41b63157-bdec-47c2-a4b0-228f89541a2a" (UID: "41b63157-bdec-47c2-a4b0-228f89541a2a"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.445020 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "41b63157-bdec-47c2-a4b0-228f89541a2a" (UID: "41b63157-bdec-47c2-a4b0-228f89541a2a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.454489 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41b63157-bdec-47c2-a4b0-228f89541a2a-kube-api-access-rbgw2" (OuterVolumeSpecName: "kube-api-access-rbgw2") pod "41b63157-bdec-47c2-a4b0-228f89541a2a" (UID: "41b63157-bdec-47c2-a4b0-228f89541a2a"). InnerVolumeSpecName "kube-api-access-rbgw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.461973 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "mysql-db") pod "41b63157-bdec-47c2-a4b0-228f89541a2a" (UID: "41b63157-bdec-47c2-a4b0-228f89541a2a"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.488992 4706 generic.go:334] "Generic (PLEG): container finished" podID="d5dce085-c674-49c0-883c-d1d788c1b70d" containerID="926588293cf08b116811d7ec7631f2673193207058f90717ede178a9be250f36" exitCode=1 Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.489057 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" event={"ID":"d5dce085-c674-49c0-883c-d1d788c1b70d","Type":"ContainerDied","Data":"926588293cf08b116811d7ec7631f2673193207058f90717ede178a9be250f36"} Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.489094 4706 scope.go:117] "RemoveContainer" containerID="3bed485b02f207bacdc08706926eb5cc6e3405abfc5bd01d95acd6b174d2c678" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.489666 4706 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" secret="" err="secret \"galera-openstack-dockercfg-4j5n7\" not found" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.489705 4706 scope.go:117] "RemoveContainer" containerID="926588293cf08b116811d7ec7631f2673193207058f90717ede178a9be250f36" Dec 06 14:16:19 crc kubenswrapper[4706]: E1206 14:16:19.490050 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=keystoneec25-account-delete-47t8b_cinder-kuttl-tests(d5dce085-c674-49c0-883c-d1d788c1b70d)\"" pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" podUID="d5dce085-c674-49c0-883c-d1d788c1b70d" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.492950 4706 generic.go:334] "Generic (PLEG): container finished" podID="41b63157-bdec-47c2-a4b0-228f89541a2a" containerID="56494e751042f56606058af3e4fca29a89a08904e452133514ddd5ed3269298a" exitCode=0 Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.492995 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"41b63157-bdec-47c2-a4b0-228f89541a2a","Type":"ContainerDied","Data":"56494e751042f56606058af3e4fca29a89a08904e452133514ddd5ed3269298a"} Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.493015 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"41b63157-bdec-47c2-a4b0-228f89541a2a","Type":"ContainerDied","Data":"c4c34b6ec3c3851b309aac8ca34f4844b412945e3c33c811a238d93cf71a30e0"} Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.493069 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.500558 4706 generic.go:334] "Generic (PLEG): container finished" podID="76a993fd-482d-4efd-938a-2014dda4aca1" containerID="9f53871a8284673c6d9414d0374c0bf6afe88dd86fc2f4264ab391d6f28557f2" exitCode=0 Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.501048 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"76a993fd-482d-4efd-938a-2014dda4aca1","Type":"ContainerDied","Data":"9f53871a8284673c6d9414d0374c0bf6afe88dd86fc2f4264ab391d6f28557f2"} Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.530374 4706 scope.go:117] "RemoveContainer" containerID="56494e751042f56606058af3e4fca29a89a08904e452133514ddd5ed3269298a" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.548947 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.556273 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbgw2\" (UniqueName: \"kubernetes.io/projected/41b63157-bdec-47c2-a4b0-228f89541a2a-kube-api-access-rbgw2\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.556337 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.556359 4706 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.556388 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/41b63157-bdec-47c2-a4b0-228f89541a2a-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.556406 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/41b63157-bdec-47c2-a4b0-228f89541a2a-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.556447 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.559846 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/rabbitmq-server-0" podUID="23f8c907-7a56-4da3-aefb-929a43497eb2" containerName="rabbitmq" containerID="cri-o://d659020ec94512a5441f62fd7acc50ff9f16952d2c086bd354c4ef2ed27c7465" gracePeriod=604800 Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.565640 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.578524 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.586226 4706 scope.go:117] "RemoveContainer" containerID="9553942fd0f85b5353637d9514ac2d86b968c0a0a74f85f4a90194f943c222fb" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.658133 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.658932 4706 scope.go:117] "RemoveContainer" containerID="56494e751042f56606058af3e4fca29a89a08904e452133514ddd5ed3269298a" Dec 06 14:16:19 crc kubenswrapper[4706]: E1206 14:16:19.659959 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56494e751042f56606058af3e4fca29a89a08904e452133514ddd5ed3269298a\": container with ID starting with 56494e751042f56606058af3e4fca29a89a08904e452133514ddd5ed3269298a not found: ID does not exist" containerID="56494e751042f56606058af3e4fca29a89a08904e452133514ddd5ed3269298a" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.659992 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56494e751042f56606058af3e4fca29a89a08904e452133514ddd5ed3269298a"} err="failed to get container status \"56494e751042f56606058af3e4fca29a89a08904e452133514ddd5ed3269298a\": rpc error: code = NotFound desc = could not find container \"56494e751042f56606058af3e4fca29a89a08904e452133514ddd5ed3269298a\": container with ID starting with 56494e751042f56606058af3e4fca29a89a08904e452133514ddd5ed3269298a not found: ID does not exist" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.660017 4706 scope.go:117] "RemoveContainer" containerID="9553942fd0f85b5353637d9514ac2d86b968c0a0a74f85f4a90194f943c222fb" Dec 06 14:16:19 crc kubenswrapper[4706]: E1206 14:16:19.660469 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9553942fd0f85b5353637d9514ac2d86b968c0a0a74f85f4a90194f943c222fb\": container with ID starting with 9553942fd0f85b5353637d9514ac2d86b968c0a0a74f85f4a90194f943c222fb not found: ID does not exist" containerID="9553942fd0f85b5353637d9514ac2d86b968c0a0a74f85f4a90194f943c222fb" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.660524 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9553942fd0f85b5353637d9514ac2d86b968c0a0a74f85f4a90194f943c222fb"} err="failed to get container status \"9553942fd0f85b5353637d9514ac2d86b968c0a0a74f85f4a90194f943c222fb\": rpc error: code = NotFound desc = could not find container \"9553942fd0f85b5353637d9514ac2d86b968c0a0a74f85f4a90194f943c222fb\": container with ID starting with 9553942fd0f85b5353637d9514ac2d86b968c0a0a74f85f4a90194f943c222fb not found: ID does not exist" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.686623 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.866605 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/76a993fd-482d-4efd-938a-2014dda4aca1-kolla-config\") pod \"76a993fd-482d-4efd-938a-2014dda4aca1\" (UID: \"76a993fd-482d-4efd-938a-2014dda4aca1\") " Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.866735 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76a993fd-482d-4efd-938a-2014dda4aca1-config-data\") pod \"76a993fd-482d-4efd-938a-2014dda4aca1\" (UID: \"76a993fd-482d-4efd-938a-2014dda4aca1\") " Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.866799 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vvqb\" (UniqueName: \"kubernetes.io/projected/76a993fd-482d-4efd-938a-2014dda4aca1-kube-api-access-5vvqb\") pod \"76a993fd-482d-4efd-938a-2014dda4aca1\" (UID: \"76a993fd-482d-4efd-938a-2014dda4aca1\") " Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.867212 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76a993fd-482d-4efd-938a-2014dda4aca1-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "76a993fd-482d-4efd-938a-2014dda4aca1" (UID: "76a993fd-482d-4efd-938a-2014dda4aca1"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.868110 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76a993fd-482d-4efd-938a-2014dda4aca1-config-data" (OuterVolumeSpecName: "config-data") pod "76a993fd-482d-4efd-938a-2014dda4aca1" (UID: "76a993fd-482d-4efd-938a-2014dda4aca1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.871410 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76a993fd-482d-4efd-938a-2014dda4aca1-kube-api-access-5vvqb" (OuterVolumeSpecName: "kube-api-access-5vvqb") pod "76a993fd-482d-4efd-938a-2014dda4aca1" (UID: "76a993fd-482d-4efd-938a-2014dda4aca1"). InnerVolumeSpecName "kube-api-access-5vvqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.967971 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76a993fd-482d-4efd-938a-2014dda4aca1-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.968025 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vvqb\" (UniqueName: \"kubernetes.io/projected/76a993fd-482d-4efd-938a-2014dda4aca1-kube-api-access-5vvqb\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:19 crc kubenswrapper[4706]: I1206 14:16:19.968048 4706 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/76a993fd-482d-4efd-938a-2014dda4aca1-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.016963 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/openstack-galera-1" podUID="252d778a-eb1a-4519-9325-e79910544f1d" containerName="galera" containerID="cri-o://a92d440a79be04659f5deb197d5547b6bdfbb3797c5ad7c4659c397e557c4cf9" gracePeriod=28 Dec 06 14:16:20 crc kubenswrapper[4706]: E1206 14:16:20.171944 4706 configmap.go:193] Couldn't get configMap cinder-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Dec 06 14:16:20 crc kubenswrapper[4706]: E1206 14:16:20.172048 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d5dce085-c674-49c0-883c-d1d788c1b70d-operator-scripts podName:d5dce085-c674-49c0-883c-d1d788c1b70d nodeName:}" failed. No retries permitted until 2025-12-06 14:16:22.172029699 +0000 UTC m=+1225.007822576 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/d5dce085-c674-49c0-883c-d1d788c1b70d-operator-scripts") pod "keystoneec25-account-delete-47t8b" (UID: "d5dce085-c674-49c0-883c-d1d788c1b70d") : configmap "openstack-scripts" not found Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.264990 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7"] Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.265401 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" podUID="8c97961d-a743-4c7d-9bde-f319fea2af90" containerName="manager" containerID="cri-o://ee5b6933fe042a50a48e3b44db8f21a33564dfde007d5a5b105018d966ac9cae" gracePeriod=10 Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.535527 4706 generic.go:334] "Generic (PLEG): container finished" podID="4ddad38d-0652-48ab-a70e-7eca376ba89d" containerID="22412e6cb2fd24e10cf3f94722611789ed23e1b4208f9601d256eca63c15b73c" exitCode=0 Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.535607 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" event={"ID":"4ddad38d-0652-48ab-a70e-7eca376ba89d","Type":"ContainerDied","Data":"22412e6cb2fd24e10cf3f94722611789ed23e1b4208f9601d256eca63c15b73c"} Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.576061 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/cinder-operator-index-klwkp"] Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.576125 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"76a993fd-482d-4efd-938a-2014dda4aca1","Type":"ContainerDied","Data":"0c6aef584ebedf1f371a4af402cc6896b08e5195e233691a8faf4830ef5705dc"} Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.576164 4706 scope.go:117] "RemoveContainer" containerID="9f53871a8284673c6d9414d0374c0bf6afe88dd86fc2f4264ab391d6f28557f2" Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.576187 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.586185 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/cinder-operator-index-klwkp" podUID="c36a0ce1-c1a6-4529-9579-a48508a688fc" containerName="registry-server" containerID="cri-o://27674d1e2788e70273f60a460cc1e0d92991def0d648c3ca66dbcab803e71fdd" gracePeriod=30 Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.591267 4706 generic.go:334] "Generic (PLEG): container finished" podID="8c97961d-a743-4c7d-9bde-f319fea2af90" containerID="ee5b6933fe042a50a48e3b44db8f21a33564dfde007d5a5b105018d966ac9cae" exitCode=0 Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.591323 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" event={"ID":"8c97961d-a743-4c7d-9bde-f319fea2af90","Type":"ContainerDied","Data":"ee5b6933fe042a50a48e3b44db8f21a33564dfde007d5a5b105018d966ac9cae"} Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.610982 4706 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" secret="" err="secret \"galera-openstack-dockercfg-4j5n7\" not found" Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.611052 4706 scope.go:117] "RemoveContainer" containerID="926588293cf08b116811d7ec7631f2673193207058f90717ede178a9be250f36" Dec 06 14:16:20 crc kubenswrapper[4706]: E1206 14:16:20.611207 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=keystoneec25-account-delete-47t8b_cinder-kuttl-tests(d5dce085-c674-49c0-883c-d1d788c1b70d)\"" pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" podUID="d5dce085-c674-49c0-883c-d1d788c1b70d" Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.645628 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf"] Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.691905 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/be6924214bfadb9475308df976d961458bdcf9028f98c3f594f4c1ca66h4dsf"] Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.847008 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.858490 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.862344 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.865189 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.908205 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-fernet-keys\") pod \"4ddad38d-0652-48ab-a70e-7eca376ba89d\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.908247 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-config-data\") pod \"4ddad38d-0652-48ab-a70e-7eca376ba89d\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.908271 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8c97961d-a743-4c7d-9bde-f319fea2af90-apiservice-cert\") pod \"8c97961d-a743-4c7d-9bde-f319fea2af90\" (UID: \"8c97961d-a743-4c7d-9bde-f319fea2af90\") " Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.908323 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-scripts\") pod \"4ddad38d-0652-48ab-a70e-7eca376ba89d\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.908353 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbg72\" (UniqueName: \"kubernetes.io/projected/4ddad38d-0652-48ab-a70e-7eca376ba89d-kube-api-access-qbg72\") pod \"4ddad38d-0652-48ab-a70e-7eca376ba89d\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.908392 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8c97961d-a743-4c7d-9bde-f319fea2af90-webhook-cert\") pod \"8c97961d-a743-4c7d-9bde-f319fea2af90\" (UID: \"8c97961d-a743-4c7d-9bde-f319fea2af90\") " Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.908406 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rp5x8\" (UniqueName: \"kubernetes.io/projected/8c97961d-a743-4c7d-9bde-f319fea2af90-kube-api-access-rp5x8\") pod \"8c97961d-a743-4c7d-9bde-f319fea2af90\" (UID: \"8c97961d-a743-4c7d-9bde-f319fea2af90\") " Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.908427 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-credential-keys\") pod \"4ddad38d-0652-48ab-a70e-7eca376ba89d\" (UID: \"4ddad38d-0652-48ab-a70e-7eca376ba89d\") " Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.923106 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c97961d-a743-4c7d-9bde-f319fea2af90-kube-api-access-rp5x8" (OuterVolumeSpecName: "kube-api-access-rp5x8") pod "8c97961d-a743-4c7d-9bde-f319fea2af90" (UID: "8c97961d-a743-4c7d-9bde-f319fea2af90"). InnerVolumeSpecName "kube-api-access-rp5x8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.923637 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-scripts" (OuterVolumeSpecName: "scripts") pod "4ddad38d-0652-48ab-a70e-7eca376ba89d" (UID: "4ddad38d-0652-48ab-a70e-7eca376ba89d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.924417 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4ddad38d-0652-48ab-a70e-7eca376ba89d" (UID: "4ddad38d-0652-48ab-a70e-7eca376ba89d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.926406 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c97961d-a743-4c7d-9bde-f319fea2af90-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "8c97961d-a743-4c7d-9bde-f319fea2af90" (UID: "8c97961d-a743-4c7d-9bde-f319fea2af90"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.926538 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c97961d-a743-4c7d-9bde-f319fea2af90-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "8c97961d-a743-4c7d-9bde-f319fea2af90" (UID: "8c97961d-a743-4c7d-9bde-f319fea2af90"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.926571 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ddad38d-0652-48ab-a70e-7eca376ba89d-kube-api-access-qbg72" (OuterVolumeSpecName: "kube-api-access-qbg72") pod "4ddad38d-0652-48ab-a70e-7eca376ba89d" (UID: "4ddad38d-0652-48ab-a70e-7eca376ba89d"). InnerVolumeSpecName "kube-api-access-qbg72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.928470 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4ddad38d-0652-48ab-a70e-7eca376ba89d" (UID: "4ddad38d-0652-48ab-a70e-7eca376ba89d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.941417 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-config-data" (OuterVolumeSpecName: "config-data") pod "4ddad38d-0652-48ab-a70e-7eca376ba89d" (UID: "4ddad38d-0652-48ab-a70e-7eca376ba89d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:20 crc kubenswrapper[4706]: I1206 14:16:20.997527 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-klwkp" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.009413 4706 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.009444 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.009458 4706 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8c97961d-a743-4c7d-9bde-f319fea2af90-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.009472 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.009484 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbg72\" (UniqueName: \"kubernetes.io/projected/4ddad38d-0652-48ab-a70e-7eca376ba89d-kube-api-access-qbg72\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.009496 4706 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8c97961d-a743-4c7d-9bde-f319fea2af90-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.009507 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rp5x8\" (UniqueName: \"kubernetes.io/projected/8c97961d-a743-4c7d-9bde-f319fea2af90-kube-api-access-rp5x8\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.009518 4706 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4ddad38d-0652-48ab-a70e-7eca376ba89d-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.092203 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.110570 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zt9vs\" (UniqueName: \"kubernetes.io/projected/c36a0ce1-c1a6-4529-9579-a48508a688fc-kube-api-access-zt9vs\") pod \"c36a0ce1-c1a6-4529-9579-a48508a688fc\" (UID: \"c36a0ce1-c1a6-4529-9579-a48508a688fc\") " Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.114801 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c36a0ce1-c1a6-4529-9579-a48508a688fc-kube-api-access-zt9vs" (OuterVolumeSpecName: "kube-api-access-zt9vs") pod "c36a0ce1-c1a6-4529-9579-a48508a688fc" (UID: "c36a0ce1-c1a6-4529-9579-a48508a688fc"). InnerVolumeSpecName "kube-api-access-zt9vs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.211604 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85dmj\" (UniqueName: \"kubernetes.io/projected/23f8c907-7a56-4da3-aefb-929a43497eb2-kube-api-access-85dmj\") pod \"23f8c907-7a56-4da3-aefb-929a43497eb2\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.211647 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/23f8c907-7a56-4da3-aefb-929a43497eb2-pod-info\") pod \"23f8c907-7a56-4da3-aefb-929a43497eb2\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.211697 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/23f8c907-7a56-4da3-aefb-929a43497eb2-erlang-cookie-secret\") pod \"23f8c907-7a56-4da3-aefb-929a43497eb2\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.211806 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a988e02-edca-43b9-b5e5-92410a885971\") pod \"23f8c907-7a56-4da3-aefb-929a43497eb2\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.211851 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-erlang-cookie\") pod \"23f8c907-7a56-4da3-aefb-929a43497eb2\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.211871 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-plugins\") pod \"23f8c907-7a56-4da3-aefb-929a43497eb2\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.211891 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-confd\") pod \"23f8c907-7a56-4da3-aefb-929a43497eb2\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.211915 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/23f8c907-7a56-4da3-aefb-929a43497eb2-plugins-conf\") pod \"23f8c907-7a56-4da3-aefb-929a43497eb2\" (UID: \"23f8c907-7a56-4da3-aefb-929a43497eb2\") " Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.212174 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zt9vs\" (UniqueName: \"kubernetes.io/projected/c36a0ce1-c1a6-4529-9579-a48508a688fc-kube-api-access-zt9vs\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.213058 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "23f8c907-7a56-4da3-aefb-929a43497eb2" (UID: "23f8c907-7a56-4da3-aefb-929a43497eb2"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.213075 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "23f8c907-7a56-4da3-aefb-929a43497eb2" (UID: "23f8c907-7a56-4da3-aefb-929a43497eb2"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.213523 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23f8c907-7a56-4da3-aefb-929a43497eb2-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "23f8c907-7a56-4da3-aefb-929a43497eb2" (UID: "23f8c907-7a56-4da3-aefb-929a43497eb2"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.216443 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/23f8c907-7a56-4da3-aefb-929a43497eb2-pod-info" (OuterVolumeSpecName: "pod-info") pod "23f8c907-7a56-4da3-aefb-929a43497eb2" (UID: "23f8c907-7a56-4da3-aefb-929a43497eb2"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.216462 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23f8c907-7a56-4da3-aefb-929a43497eb2-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "23f8c907-7a56-4da3-aefb-929a43497eb2" (UID: "23f8c907-7a56-4da3-aefb-929a43497eb2"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.217421 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23f8c907-7a56-4da3-aefb-929a43497eb2-kube-api-access-85dmj" (OuterVolumeSpecName: "kube-api-access-85dmj") pod "23f8c907-7a56-4da3-aefb-929a43497eb2" (UID: "23f8c907-7a56-4da3-aefb-929a43497eb2"). InnerVolumeSpecName "kube-api-access-85dmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.220259 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a988e02-edca-43b9-b5e5-92410a885971" (OuterVolumeSpecName: "persistence") pod "23f8c907-7a56-4da3-aefb-929a43497eb2" (UID: "23f8c907-7a56-4da3-aefb-929a43497eb2"). InnerVolumeSpecName "pvc-1a988e02-edca-43b9-b5e5-92410a885971". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.282917 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "23f8c907-7a56-4da3-aefb-929a43497eb2" (UID: "23f8c907-7a56-4da3-aefb-929a43497eb2"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.313719 4706 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/23f8c907-7a56-4da3-aefb-929a43497eb2-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.313842 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-1a988e02-edca-43b9-b5e5-92410a885971\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a988e02-edca-43b9-b5e5-92410a885971\") on node \"crc\" " Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.313886 4706 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.313917 4706 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.313945 4706 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/23f8c907-7a56-4da3-aefb-929a43497eb2-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.313970 4706 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/23f8c907-7a56-4da3-aefb-929a43497eb2-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.313995 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85dmj\" (UniqueName: \"kubernetes.io/projected/23f8c907-7a56-4da3-aefb-929a43497eb2-kube-api-access-85dmj\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.314025 4706 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/23f8c907-7a56-4da3-aefb-929a43497eb2-pod-info\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.327027 4706 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.327184 4706 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-1a988e02-edca-43b9-b5e5-92410a885971" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a988e02-edca-43b9-b5e5-92410a885971") on node "crc" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.415178 4706 reconciler_common.go:293] "Volume detached for volume \"pvc-1a988e02-edca-43b9-b5e5-92410a885971\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a988e02-edca-43b9-b5e5-92410a885971\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.528576 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41b63157-bdec-47c2-a4b0-228f89541a2a" path="/var/lib/kubelet/pods/41b63157-bdec-47c2-a4b0-228f89541a2a/volumes" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.536379 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76a993fd-482d-4efd-938a-2014dda4aca1" path="/var/lib/kubelet/pods/76a993fd-482d-4efd-938a-2014dda4aca1/volumes" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.537177 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96f74c30-7693-4aa2-a171-5739c79cee2a" path="/var/lib/kubelet/pods/96f74c30-7693-4aa2-a171-5739c79cee2a/volumes" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.620983 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" event={"ID":"8c97961d-a743-4c7d-9bde-f319fea2af90","Type":"ContainerDied","Data":"d26e4af2db8eae4bfd2383991852132847f374ce781c66bf03650a0e5abe3b94"} Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.621048 4706 scope.go:117] "RemoveContainer" containerID="ee5b6933fe042a50a48e3b44db8f21a33564dfde007d5a5b105018d966ac9cae" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.621167 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.643650 4706 generic.go:334] "Generic (PLEG): container finished" podID="c36a0ce1-c1a6-4529-9579-a48508a688fc" containerID="27674d1e2788e70273f60a460cc1e0d92991def0d648c3ca66dbcab803e71fdd" exitCode=0 Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.643690 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-klwkp" event={"ID":"c36a0ce1-c1a6-4529-9579-a48508a688fc","Type":"ContainerDied","Data":"27674d1e2788e70273f60a460cc1e0d92991def0d648c3ca66dbcab803e71fdd"} Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.643707 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-klwkp" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.643724 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-klwkp" event={"ID":"c36a0ce1-c1a6-4529-9579-a48508a688fc","Type":"ContainerDied","Data":"c9aa9b2e3f6b74c0d3f67dd38ffdc10e89e7f098e277eceb3edba07a98854646"} Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.651373 4706 generic.go:334] "Generic (PLEG): container finished" podID="23f8c907-7a56-4da3-aefb-929a43497eb2" containerID="d659020ec94512a5441f62fd7acc50ff9f16952d2c086bd354c4ef2ed27c7465" exitCode=0 Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.651456 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"23f8c907-7a56-4da3-aefb-929a43497eb2","Type":"ContainerDied","Data":"d659020ec94512a5441f62fd7acc50ff9f16952d2c086bd354c4ef2ed27c7465"} Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.651490 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"23f8c907-7a56-4da3-aefb-929a43497eb2","Type":"ContainerDied","Data":"958295388ed3710103707281add52f4b201f078a5e63d4377f3512e64b08714f"} Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.651559 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.656659 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" event={"ID":"4ddad38d-0652-48ab-a70e-7eca376ba89d","Type":"ContainerDied","Data":"78cd8cb0e29c83e185cb9381c70f13284fb64e8840537dff26c90a1d2c228e05"} Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.656781 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.711572 4706 scope.go:117] "RemoveContainer" containerID="27674d1e2788e70273f60a460cc1e0d92991def0d648c3ca66dbcab803e71fdd" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.717244 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.741185 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.741227 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/cinder-operator-index-klwkp"] Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.747135 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/cinder-operator-index-klwkp"] Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.748359 4706 scope.go:117] "RemoveContainer" containerID="27674d1e2788e70273f60a460cc1e0d92991def0d648c3ca66dbcab803e71fdd" Dec 06 14:16:21 crc kubenswrapper[4706]: E1206 14:16:21.748880 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27674d1e2788e70273f60a460cc1e0d92991def0d648c3ca66dbcab803e71fdd\": container with ID starting with 27674d1e2788e70273f60a460cc1e0d92991def0d648c3ca66dbcab803e71fdd not found: ID does not exist" containerID="27674d1e2788e70273f60a460cc1e0d92991def0d648c3ca66dbcab803e71fdd" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.748921 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27674d1e2788e70273f60a460cc1e0d92991def0d648c3ca66dbcab803e71fdd"} err="failed to get container status \"27674d1e2788e70273f60a460cc1e0d92991def0d648c3ca66dbcab803e71fdd\": rpc error: code = NotFound desc = could not find container \"27674d1e2788e70273f60a460cc1e0d92991def0d648c3ca66dbcab803e71fdd\": container with ID starting with 27674d1e2788e70273f60a460cc1e0d92991def0d648c3ca66dbcab803e71fdd not found: ID does not exist" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.748955 4706 scope.go:117] "RemoveContainer" containerID="d659020ec94512a5441f62fd7acc50ff9f16952d2c086bd354c4ef2ed27c7465" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.754379 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh"] Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.759251 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-7cddcf8fd6-knkjh"] Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.770770 4706 scope.go:117] "RemoveContainer" containerID="6a2c055976a1f626f8c484f7e2eab83c12e64177e0a91754b946043503eb7531" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.773274 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7"] Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.777995 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-65b4f9dcf8-ltdm7"] Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.788396 4706 scope.go:117] "RemoveContainer" containerID="d659020ec94512a5441f62fd7acc50ff9f16952d2c086bd354c4ef2ed27c7465" Dec 06 14:16:21 crc kubenswrapper[4706]: E1206 14:16:21.788737 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d659020ec94512a5441f62fd7acc50ff9f16952d2c086bd354c4ef2ed27c7465\": container with ID starting with d659020ec94512a5441f62fd7acc50ff9f16952d2c086bd354c4ef2ed27c7465 not found: ID does not exist" containerID="d659020ec94512a5441f62fd7acc50ff9f16952d2c086bd354c4ef2ed27c7465" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.788771 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d659020ec94512a5441f62fd7acc50ff9f16952d2c086bd354c4ef2ed27c7465"} err="failed to get container status \"d659020ec94512a5441f62fd7acc50ff9f16952d2c086bd354c4ef2ed27c7465\": rpc error: code = NotFound desc = could not find container \"d659020ec94512a5441f62fd7acc50ff9f16952d2c086bd354c4ef2ed27c7465\": container with ID starting with d659020ec94512a5441f62fd7acc50ff9f16952d2c086bd354c4ef2ed27c7465 not found: ID does not exist" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.788796 4706 scope.go:117] "RemoveContainer" containerID="6a2c055976a1f626f8c484f7e2eab83c12e64177e0a91754b946043503eb7531" Dec 06 14:16:21 crc kubenswrapper[4706]: E1206 14:16:21.789046 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a2c055976a1f626f8c484f7e2eab83c12e64177e0a91754b946043503eb7531\": container with ID starting with 6a2c055976a1f626f8c484f7e2eab83c12e64177e0a91754b946043503eb7531 not found: ID does not exist" containerID="6a2c055976a1f626f8c484f7e2eab83c12e64177e0a91754b946043503eb7531" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.789072 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a2c055976a1f626f8c484f7e2eab83c12e64177e0a91754b946043503eb7531"} err="failed to get container status \"6a2c055976a1f626f8c484f7e2eab83c12e64177e0a91754b946043503eb7531\": rpc error: code = NotFound desc = could not find container \"6a2c055976a1f626f8c484f7e2eab83c12e64177e0a91754b946043503eb7531\": container with ID starting with 6a2c055976a1f626f8c484f7e2eab83c12e64177e0a91754b946043503eb7531 not found: ID does not exist" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.789090 4706 scope.go:117] "RemoveContainer" containerID="22412e6cb2fd24e10cf3f94722611789ed23e1b4208f9601d256eca63c15b73c" Dec 06 14:16:21 crc kubenswrapper[4706]: I1206 14:16:21.919424 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.000388 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-57889"] Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.004371 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-57889"] Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.015730 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf"] Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.022463 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/252d778a-eb1a-4519-9325-e79910544f1d-config-data-generated\") pod \"252d778a-eb1a-4519-9325-e79910544f1d\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.022514 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dl96\" (UniqueName: \"kubernetes.io/projected/252d778a-eb1a-4519-9325-e79910544f1d-kube-api-access-9dl96\") pod \"252d778a-eb1a-4519-9325-e79910544f1d\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.022603 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-operator-scripts\") pod \"252d778a-eb1a-4519-9325-e79910544f1d\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.022627 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-config-data-default\") pod \"252d778a-eb1a-4519-9325-e79910544f1d\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.022650 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-kolla-config\") pod \"252d778a-eb1a-4519-9325-e79910544f1d\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.022683 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"252d778a-eb1a-4519-9325-e79910544f1d\" (UID: \"252d778a-eb1a-4519-9325-e79910544f1d\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.022997 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/252d778a-eb1a-4519-9325-e79910544f1d-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "252d778a-eb1a-4519-9325-e79910544f1d" (UID: "252d778a-eb1a-4519-9325-e79910544f1d"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.023492 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "252d778a-eb1a-4519-9325-e79910544f1d" (UID: "252d778a-eb1a-4519-9325-e79910544f1d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.023863 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "252d778a-eb1a-4519-9325-e79910544f1d" (UID: "252d778a-eb1a-4519-9325-e79910544f1d"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.025558 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "252d778a-eb1a-4519-9325-e79910544f1d" (UID: "252d778a-eb1a-4519-9325-e79910544f1d"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.027205 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/252d778a-eb1a-4519-9325-e79910544f1d-kube-api-access-9dl96" (OuterVolumeSpecName: "kube-api-access-9dl96") pod "252d778a-eb1a-4519-9325-e79910544f1d" (UID: "252d778a-eb1a-4519-9325-e79910544f1d"). InnerVolumeSpecName "kube-api-access-9dl96". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.029537 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystoneec25-account-delete-47t8b"] Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.033358 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "mysql-db") pod "252d778a-eb1a-4519-9325-e79910544f1d" (UID: "252d778a-eb1a-4519-9325-e79910544f1d"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.034998 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/openstack-galera-0" podUID="a86ee400-6a8a-4cc6-bdea-62a40954a8b2" containerName="galera" containerID="cri-o://fafab63b20f39590bf945c4153224cf7d8639a7ffad58deffc95be89683694e0" gracePeriod=26 Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.036822 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-ec25-account-create-update-7b5jf"] Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.123769 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/252d778a-eb1a-4519-9325-e79910544f1d-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.123811 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dl96\" (UniqueName: \"kubernetes.io/projected/252d778a-eb1a-4519-9325-e79910544f1d-kube-api-access-9dl96\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.123824 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.123837 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.123851 4706 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/252d778a-eb1a-4519-9325-e79910544f1d-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.123887 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.137621 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.145445 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc"] Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.145679 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" podUID="40dc67dc-8180-452e-9345-a3e52b2adc15" containerName="manager" containerID="cri-o://9cda45ed2d5c17ab5baaa830233df7136e195395cffb205681faba4a417e4a39" gracePeriod=10 Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.224740 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:22 crc kubenswrapper[4706]: E1206 14:16:22.224765 4706 configmap.go:193] Couldn't get configMap cinder-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Dec 06 14:16:22 crc kubenswrapper[4706]: E1206 14:16:22.224849 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d5dce085-c674-49c0-883c-d1d788c1b70d-operator-scripts podName:d5dce085-c674-49c0-883c-d1d788c1b70d nodeName:}" failed. No retries permitted until 2025-12-06 14:16:26.224830337 +0000 UTC m=+1229.060623194 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/d5dce085-c674-49c0-883c-d1d788c1b70d-operator-scripts") pod "keystoneec25-account-delete-47t8b" (UID: "d5dce085-c674-49c0-883c-d1d788c1b70d") : configmap "openstack-scripts" not found Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.331250 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.396691 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-z4lsq"] Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.396920 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-z4lsq" podUID="7490dd58-b444-46c5-ad33-b953af4e6617" containerName="registry-server" containerID="cri-o://e6b2578eedd50f1964b40e7247cfb412e66462b9860027502721d0a39f08c826" gracePeriod=30 Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.436734 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r"] Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.454291 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85hpn9r"] Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.529693 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5dce085-c674-49c0-883c-d1d788c1b70d-operator-scripts\") pod \"d5dce085-c674-49c0-883c-d1d788c1b70d\" (UID: \"d5dce085-c674-49c0-883c-d1d788c1b70d\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.529796 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5d6km\" (UniqueName: \"kubernetes.io/projected/d5dce085-c674-49c0-883c-d1d788c1b70d-kube-api-access-5d6km\") pod \"d5dce085-c674-49c0-883c-d1d788c1b70d\" (UID: \"d5dce085-c674-49c0-883c-d1d788c1b70d\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.530375 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5dce085-c674-49c0-883c-d1d788c1b70d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d5dce085-c674-49c0-883c-d1d788c1b70d" (UID: "d5dce085-c674-49c0-883c-d1d788c1b70d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.535756 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5dce085-c674-49c0-883c-d1d788c1b70d-kube-api-access-5d6km" (OuterVolumeSpecName: "kube-api-access-5d6km") pod "d5dce085-c674-49c0-883c-d1d788c1b70d" (UID: "d5dce085-c674-49c0-883c-d1d788c1b70d"). InnerVolumeSpecName "kube-api-access-5d6km". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.631847 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5d6km\" (UniqueName: \"kubernetes.io/projected/d5dce085-c674-49c0-883c-d1d788c1b70d-kube-api-access-5d6km\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.632199 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5dce085-c674-49c0-883c-d1d788c1b70d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.664205 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.670828 4706 generic.go:334] "Generic (PLEG): container finished" podID="252d778a-eb1a-4519-9325-e79910544f1d" containerID="a92d440a79be04659f5deb197d5547b6bdfbb3797c5ad7c4659c397e557c4cf9" exitCode=0 Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.670880 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"252d778a-eb1a-4519-9325-e79910544f1d","Type":"ContainerDied","Data":"a92d440a79be04659f5deb197d5547b6bdfbb3797c5ad7c4659c397e557c4cf9"} Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.670901 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"252d778a-eb1a-4519-9325-e79910544f1d","Type":"ContainerDied","Data":"16684f78471591fa81eefb61f91abecf5e00c05be216dd56984380415a77c37f"} Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.670919 4706 scope.go:117] "RemoveContainer" containerID="a92d440a79be04659f5deb197d5547b6bdfbb3797c5ad7c4659c397e557c4cf9" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.671014 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.673580 4706 generic.go:334] "Generic (PLEG): container finished" podID="7490dd58-b444-46c5-ad33-b953af4e6617" containerID="e6b2578eedd50f1964b40e7247cfb412e66462b9860027502721d0a39f08c826" exitCode=0 Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.673648 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-z4lsq" event={"ID":"7490dd58-b444-46c5-ad33-b953af4e6617","Type":"ContainerDied","Data":"e6b2578eedd50f1964b40e7247cfb412e66462b9860027502721d0a39f08c826"} Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.675081 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" event={"ID":"d5dce085-c674-49c0-883c-d1d788c1b70d","Type":"ContainerDied","Data":"ccdc38bcab851c50cdfc7003f71b98c26fb5f2faa16f930d5340504ba2c7a6c2"} Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.675123 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystoneec25-account-delete-47t8b" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.677063 4706 generic.go:334] "Generic (PLEG): container finished" podID="a86ee400-6a8a-4cc6-bdea-62a40954a8b2" containerID="fafab63b20f39590bf945c4153224cf7d8639a7ffad58deffc95be89683694e0" exitCode=0 Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.677114 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"a86ee400-6a8a-4cc6-bdea-62a40954a8b2","Type":"ContainerDied","Data":"fafab63b20f39590bf945c4153224cf7d8639a7ffad58deffc95be89683694e0"} Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.679226 4706 generic.go:334] "Generic (PLEG): container finished" podID="40dc67dc-8180-452e-9345-a3e52b2adc15" containerID="9cda45ed2d5c17ab5baaa830233df7136e195395cffb205681faba4a417e4a39" exitCode=0 Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.679267 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" event={"ID":"40dc67dc-8180-452e-9345-a3e52b2adc15","Type":"ContainerDied","Data":"9cda45ed2d5c17ab5baaa830233df7136e195395cffb205681faba4a417e4a39"} Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.679282 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" event={"ID":"40dc67dc-8180-452e-9345-a3e52b2adc15","Type":"ContainerDied","Data":"cccb629ebd646bea24483a7ec1acc9633a8aba1bbd60a81df8fa6e1c724339bd"} Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.679333 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.709092 4706 scope.go:117] "RemoveContainer" containerID="5041e6e52899dbad5204aedfe6bd40b7b14f8fa0b7a3bf46c4500dcc5b75f5df" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.718262 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.728242 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.734417 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystoneec25-account-delete-47t8b"] Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.736830 4706 scope.go:117] "RemoveContainer" containerID="a92d440a79be04659f5deb197d5547b6bdfbb3797c5ad7c4659c397e557c4cf9" Dec 06 14:16:22 crc kubenswrapper[4706]: E1206 14:16:22.738089 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a92d440a79be04659f5deb197d5547b6bdfbb3797c5ad7c4659c397e557c4cf9\": container with ID starting with a92d440a79be04659f5deb197d5547b6bdfbb3797c5ad7c4659c397e557c4cf9 not found: ID does not exist" containerID="a92d440a79be04659f5deb197d5547b6bdfbb3797c5ad7c4659c397e557c4cf9" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.738114 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a92d440a79be04659f5deb197d5547b6bdfbb3797c5ad7c4659c397e557c4cf9"} err="failed to get container status \"a92d440a79be04659f5deb197d5547b6bdfbb3797c5ad7c4659c397e557c4cf9\": rpc error: code = NotFound desc = could not find container \"a92d440a79be04659f5deb197d5547b6bdfbb3797c5ad7c4659c397e557c4cf9\": container with ID starting with a92d440a79be04659f5deb197d5547b6bdfbb3797c5ad7c4659c397e557c4cf9 not found: ID does not exist" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.738134 4706 scope.go:117] "RemoveContainer" containerID="5041e6e52899dbad5204aedfe6bd40b7b14f8fa0b7a3bf46c4500dcc5b75f5df" Dec 06 14:16:22 crc kubenswrapper[4706]: E1206 14:16:22.740184 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5041e6e52899dbad5204aedfe6bd40b7b14f8fa0b7a3bf46c4500dcc5b75f5df\": container with ID starting with 5041e6e52899dbad5204aedfe6bd40b7b14f8fa0b7a3bf46c4500dcc5b75f5df not found: ID does not exist" containerID="5041e6e52899dbad5204aedfe6bd40b7b14f8fa0b7a3bf46c4500dcc5b75f5df" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.740202 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5041e6e52899dbad5204aedfe6bd40b7b14f8fa0b7a3bf46c4500dcc5b75f5df"} err="failed to get container status \"5041e6e52899dbad5204aedfe6bd40b7b14f8fa0b7a3bf46c4500dcc5b75f5df\": rpc error: code = NotFound desc = could not find container \"5041e6e52899dbad5204aedfe6bd40b7b14f8fa0b7a3bf46c4500dcc5b75f5df\": container with ID starting with 5041e6e52899dbad5204aedfe6bd40b7b14f8fa0b7a3bf46c4500dcc5b75f5df not found: ID does not exist" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.740215 4706 scope.go:117] "RemoveContainer" containerID="926588293cf08b116811d7ec7631f2673193207058f90717ede178a9be250f36" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.740787 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystoneec25-account-delete-47t8b"] Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.778913 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.791094 4706 scope.go:117] "RemoveContainer" containerID="9cda45ed2d5c17ab5baaa830233df7136e195395cffb205681faba4a417e4a39" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.805028 4706 scope.go:117] "RemoveContainer" containerID="9cda45ed2d5c17ab5baaa830233df7136e195395cffb205681faba4a417e4a39" Dec 06 14:16:22 crc kubenswrapper[4706]: E1206 14:16:22.805719 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cda45ed2d5c17ab5baaa830233df7136e195395cffb205681faba4a417e4a39\": container with ID starting with 9cda45ed2d5c17ab5baaa830233df7136e195395cffb205681faba4a417e4a39 not found: ID does not exist" containerID="9cda45ed2d5c17ab5baaa830233df7136e195395cffb205681faba4a417e4a39" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.805756 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cda45ed2d5c17ab5baaa830233df7136e195395cffb205681faba4a417e4a39"} err="failed to get container status \"9cda45ed2d5c17ab5baaa830233df7136e195395cffb205681faba4a417e4a39\": rpc error: code = NotFound desc = could not find container \"9cda45ed2d5c17ab5baaa830233df7136e195395cffb205681faba4a417e4a39\": container with ID starting with 9cda45ed2d5c17ab5baaa830233df7136e195395cffb205681faba4a417e4a39 not found: ID does not exist" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.835822 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/40dc67dc-8180-452e-9345-a3e52b2adc15-apiservice-cert\") pod \"40dc67dc-8180-452e-9345-a3e52b2adc15\" (UID: \"40dc67dc-8180-452e-9345-a3e52b2adc15\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.835910 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/40dc67dc-8180-452e-9345-a3e52b2adc15-webhook-cert\") pod \"40dc67dc-8180-452e-9345-a3e52b2adc15\" (UID: \"40dc67dc-8180-452e-9345-a3e52b2adc15\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.835949 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8zp7\" (UniqueName: \"kubernetes.io/projected/40dc67dc-8180-452e-9345-a3e52b2adc15-kube-api-access-c8zp7\") pod \"40dc67dc-8180-452e-9345-a3e52b2adc15\" (UID: \"40dc67dc-8180-452e-9345-a3e52b2adc15\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.839813 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40dc67dc-8180-452e-9345-a3e52b2adc15-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "40dc67dc-8180-452e-9345-a3e52b2adc15" (UID: "40dc67dc-8180-452e-9345-a3e52b2adc15"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.839832 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40dc67dc-8180-452e-9345-a3e52b2adc15-kube-api-access-c8zp7" (OuterVolumeSpecName: "kube-api-access-c8zp7") pod "40dc67dc-8180-452e-9345-a3e52b2adc15" (UID: "40dc67dc-8180-452e-9345-a3e52b2adc15"). InnerVolumeSpecName "kube-api-access-c8zp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.840331 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40dc67dc-8180-452e-9345-a3e52b2adc15-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "40dc67dc-8180-452e-9345-a3e52b2adc15" (UID: "40dc67dc-8180-452e-9345-a3e52b2adc15"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.876150 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-z4lsq" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.937802 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-kolla-config\") pod \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.937888 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bhcc\" (UniqueName: \"kubernetes.io/projected/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-kube-api-access-9bhcc\") pod \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.937945 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-config-data-generated\") pod \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.937972 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8rw5\" (UniqueName: \"kubernetes.io/projected/7490dd58-b444-46c5-ad33-b953af4e6617-kube-api-access-p8rw5\") pod \"7490dd58-b444-46c5-ad33-b953af4e6617\" (UID: \"7490dd58-b444-46c5-ad33-b953af4e6617\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.937995 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.938060 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-operator-scripts\") pod \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.938101 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-config-data-default\") pod \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\" (UID: \"a86ee400-6a8a-4cc6-bdea-62a40954a8b2\") " Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.938296 4706 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/40dc67dc-8180-452e-9345-a3e52b2adc15-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.938334 4706 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/40dc67dc-8180-452e-9345-a3e52b2adc15-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.938348 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8zp7\" (UniqueName: \"kubernetes.io/projected/40dc67dc-8180-452e-9345-a3e52b2adc15-kube-api-access-c8zp7\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.939248 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "a86ee400-6a8a-4cc6-bdea-62a40954a8b2" (UID: "a86ee400-6a8a-4cc6-bdea-62a40954a8b2"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.939378 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "a86ee400-6a8a-4cc6-bdea-62a40954a8b2" (UID: "a86ee400-6a8a-4cc6-bdea-62a40954a8b2"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.939417 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "a86ee400-6a8a-4cc6-bdea-62a40954a8b2" (UID: "a86ee400-6a8a-4cc6-bdea-62a40954a8b2"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.940072 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a86ee400-6a8a-4cc6-bdea-62a40954a8b2" (UID: "a86ee400-6a8a-4cc6-bdea-62a40954a8b2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.941364 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-kube-api-access-9bhcc" (OuterVolumeSpecName: "kube-api-access-9bhcc") pod "a86ee400-6a8a-4cc6-bdea-62a40954a8b2" (UID: "a86ee400-6a8a-4cc6-bdea-62a40954a8b2"). InnerVolumeSpecName "kube-api-access-9bhcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.941540 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7490dd58-b444-46c5-ad33-b953af4e6617-kube-api-access-p8rw5" (OuterVolumeSpecName: "kube-api-access-p8rw5") pod "7490dd58-b444-46c5-ad33-b953af4e6617" (UID: "7490dd58-b444-46c5-ad33-b953af4e6617"). InnerVolumeSpecName "kube-api-access-p8rw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:22 crc kubenswrapper[4706]: I1206 14:16:22.946506 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "mysql-db") pod "a86ee400-6a8a-4cc6-bdea-62a40954a8b2" (UID: "a86ee400-6a8a-4cc6-bdea-62a40954a8b2"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.010078 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc"] Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.013846 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-85d99cc6c-dszkc"] Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.039003 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.039038 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.039048 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.039057 4706 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.039066 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bhcc\" (UniqueName: \"kubernetes.io/projected/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-kube-api-access-9bhcc\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.039075 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a86ee400-6a8a-4cc6-bdea-62a40954a8b2-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.039085 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8rw5\" (UniqueName: \"kubernetes.io/projected/7490dd58-b444-46c5-ad33-b953af4e6617-kube-api-access-p8rw5\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.048988 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.139758 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.524253 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1774dcba-87ac-4217-9cde-29c8ef195974" path="/var/lib/kubelet/pods/1774dcba-87ac-4217-9cde-29c8ef195974/volumes" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.525019 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23f8c907-7a56-4da3-aefb-929a43497eb2" path="/var/lib/kubelet/pods/23f8c907-7a56-4da3-aefb-929a43497eb2/volumes" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.525610 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="252d778a-eb1a-4519-9325-e79910544f1d" path="/var/lib/kubelet/pods/252d778a-eb1a-4519-9325-e79910544f1d/volumes" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.526094 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40779db0-e757-45d1-956f-bc17fb6b2c62" path="/var/lib/kubelet/pods/40779db0-e757-45d1-956f-bc17fb6b2c62/volumes" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.526545 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40dc67dc-8180-452e-9345-a3e52b2adc15" path="/var/lib/kubelet/pods/40dc67dc-8180-452e-9345-a3e52b2adc15/volumes" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.526963 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ddad38d-0652-48ab-a70e-7eca376ba89d" path="/var/lib/kubelet/pods/4ddad38d-0652-48ab-a70e-7eca376ba89d/volumes" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.528295 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0" path="/var/lib/kubelet/pods/7ac35e8f-d1fe-472b-8f1d-9d0424fa6ed0/volumes" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.528794 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c97961d-a743-4c7d-9bde-f319fea2af90" path="/var/lib/kubelet/pods/8c97961d-a743-4c7d-9bde-f319fea2af90/volumes" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.529221 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c36a0ce1-c1a6-4529-9579-a48508a688fc" path="/var/lib/kubelet/pods/c36a0ce1-c1a6-4529-9579-a48508a688fc/volumes" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.529758 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5dce085-c674-49c0-883c-d1d788c1b70d" path="/var/lib/kubelet/pods/d5dce085-c674-49c0-883c-d1d788c1b70d/volumes" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.699355 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-z4lsq" event={"ID":"7490dd58-b444-46c5-ad33-b953af4e6617","Type":"ContainerDied","Data":"c40218a5fa31108baaef985e772ec444035a71eef14ef91a45e74bceca360f1f"} Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.699667 4706 scope.go:117] "RemoveContainer" containerID="e6b2578eedd50f1964b40e7247cfb412e66462b9860027502721d0a39f08c826" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.699807 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-z4lsq" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.704036 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"a86ee400-6a8a-4cc6-bdea-62a40954a8b2","Type":"ContainerDied","Data":"4db16c5d9b40a1c1ab2ba870e2692f4c045b4ddc9325d308503a5ee334dcf9f5"} Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.704253 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.728639 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-z4lsq"] Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.735581 4706 scope.go:117] "RemoveContainer" containerID="fafab63b20f39590bf945c4153224cf7d8639a7ffad58deffc95be89683694e0" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.746080 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-z4lsq"] Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.749967 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.753975 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.757363 4706 scope.go:117] "RemoveContainer" containerID="a591760eadcadf8809da63c6dd9d5afaaf5320cb90aa24fcd946f35cb81242ae" Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.930787 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp"] Dec 06 14:16:23 crc kubenswrapper[4706]: I1206 14:16:23.931022 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp" podUID="18cc7358-0421-415c-a37b-6dba9a764d93" containerName="operator" containerID="cri-o://1779e8484d779e98a1cc93c6c5788c8cb6a4a5645fb3ac2d27b558da926da210" gracePeriod=10 Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.229126 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-6xsvs"] Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.229652 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-6xsvs" podUID="352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9" containerName="registry-server" containerID="cri-o://5f39d0295437686b34ff846dc9258109e8568381c6a48e7cb5a202f60d83f2c6" gracePeriod=30 Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.248872 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx"] Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.252639 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908frwx"] Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.385564 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp" Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.563321 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k62b8\" (UniqueName: \"kubernetes.io/projected/18cc7358-0421-415c-a37b-6dba9a764d93-kube-api-access-k62b8\") pod \"18cc7358-0421-415c-a37b-6dba9a764d93\" (UID: \"18cc7358-0421-415c-a37b-6dba9a764d93\") " Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.569523 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18cc7358-0421-415c-a37b-6dba9a764d93-kube-api-access-k62b8" (OuterVolumeSpecName: "kube-api-access-k62b8") pod "18cc7358-0421-415c-a37b-6dba9a764d93" (UID: "18cc7358-0421-415c-a37b-6dba9a764d93"). InnerVolumeSpecName "kube-api-access-k62b8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.664521 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k62b8\" (UniqueName: \"kubernetes.io/projected/18cc7358-0421-415c-a37b-6dba9a764d93-kube-api-access-k62b8\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.671004 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-6xsvs" Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.722013 4706 generic.go:334] "Generic (PLEG): container finished" podID="18cc7358-0421-415c-a37b-6dba9a764d93" containerID="1779e8484d779e98a1cc93c6c5788c8cb6a4a5645fb3ac2d27b558da926da210" exitCode=0 Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.722101 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp" event={"ID":"18cc7358-0421-415c-a37b-6dba9a764d93","Type":"ContainerDied","Data":"1779e8484d779e98a1cc93c6c5788c8cb6a4a5645fb3ac2d27b558da926da210"} Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.722170 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp" event={"ID":"18cc7358-0421-415c-a37b-6dba9a764d93","Type":"ContainerDied","Data":"e9e2f744285d2bf82711a97c6c946ac701f978c36ffd90210a9d751dbffd9d38"} Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.722193 4706 scope.go:117] "RemoveContainer" containerID="1779e8484d779e98a1cc93c6c5788c8cb6a4a5645fb3ac2d27b558da926da210" Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.722129 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp" Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.728819 4706 generic.go:334] "Generic (PLEG): container finished" podID="352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9" containerID="5f39d0295437686b34ff846dc9258109e8568381c6a48e7cb5a202f60d83f2c6" exitCode=0 Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.728872 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-6xsvs" Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.728916 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-6xsvs" event={"ID":"352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9","Type":"ContainerDied","Data":"5f39d0295437686b34ff846dc9258109e8568381c6a48e7cb5a202f60d83f2c6"} Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.728953 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-6xsvs" event={"ID":"352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9","Type":"ContainerDied","Data":"ba7e718b08d89cc9658005cba4c6f43ef4e18764de6a25825d7eb3c55f5dc0d9"} Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.751349 4706 scope.go:117] "RemoveContainer" containerID="1779e8484d779e98a1cc93c6c5788c8cb6a4a5645fb3ac2d27b558da926da210" Dec 06 14:16:24 crc kubenswrapper[4706]: E1206 14:16:24.751766 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1779e8484d779e98a1cc93c6c5788c8cb6a4a5645fb3ac2d27b558da926da210\": container with ID starting with 1779e8484d779e98a1cc93c6c5788c8cb6a4a5645fb3ac2d27b558da926da210 not found: ID does not exist" containerID="1779e8484d779e98a1cc93c6c5788c8cb6a4a5645fb3ac2d27b558da926da210" Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.751803 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1779e8484d779e98a1cc93c6c5788c8cb6a4a5645fb3ac2d27b558da926da210"} err="failed to get container status \"1779e8484d779e98a1cc93c6c5788c8cb6a4a5645fb3ac2d27b558da926da210\": rpc error: code = NotFound desc = could not find container \"1779e8484d779e98a1cc93c6c5788c8cb6a4a5645fb3ac2d27b558da926da210\": container with ID starting with 1779e8484d779e98a1cc93c6c5788c8cb6a4a5645fb3ac2d27b558da926da210 not found: ID does not exist" Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.751828 4706 scope.go:117] "RemoveContainer" containerID="5f39d0295437686b34ff846dc9258109e8568381c6a48e7cb5a202f60d83f2c6" Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.752104 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp"] Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.756856 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-pqwvp"] Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.765032 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2lj6\" (UniqueName: \"kubernetes.io/projected/352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9-kube-api-access-j2lj6\") pod \"352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9\" (UID: \"352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9\") " Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.767710 4706 scope.go:117] "RemoveContainer" containerID="5f39d0295437686b34ff846dc9258109e8568381c6a48e7cb5a202f60d83f2c6" Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.767767 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9-kube-api-access-j2lj6" (OuterVolumeSpecName: "kube-api-access-j2lj6") pod "352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9" (UID: "352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9"). InnerVolumeSpecName "kube-api-access-j2lj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:24 crc kubenswrapper[4706]: E1206 14:16:24.768340 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f39d0295437686b34ff846dc9258109e8568381c6a48e7cb5a202f60d83f2c6\": container with ID starting with 5f39d0295437686b34ff846dc9258109e8568381c6a48e7cb5a202f60d83f2c6 not found: ID does not exist" containerID="5f39d0295437686b34ff846dc9258109e8568381c6a48e7cb5a202f60d83f2c6" Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.768385 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f39d0295437686b34ff846dc9258109e8568381c6a48e7cb5a202f60d83f2c6"} err="failed to get container status \"5f39d0295437686b34ff846dc9258109e8568381c6a48e7cb5a202f60d83f2c6\": rpc error: code = NotFound desc = could not find container \"5f39d0295437686b34ff846dc9258109e8568381c6a48e7cb5a202f60d83f2c6\": container with ID starting with 5f39d0295437686b34ff846dc9258109e8568381c6a48e7cb5a202f60d83f2c6 not found: ID does not exist" Dec 06 14:16:24 crc kubenswrapper[4706]: I1206 14:16:24.865852 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2lj6\" (UniqueName: \"kubernetes.io/projected/352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9-kube-api-access-j2lj6\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:25 crc kubenswrapper[4706]: I1206 14:16:25.066153 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-6xsvs"] Dec 06 14:16:25 crc kubenswrapper[4706]: I1206 14:16:25.071422 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-6xsvs"] Dec 06 14:16:25 crc kubenswrapper[4706]: I1206 14:16:25.520190 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18cc7358-0421-415c-a37b-6dba9a764d93" path="/var/lib/kubelet/pods/18cc7358-0421-415c-a37b-6dba9a764d93/volumes" Dec 06 14:16:25 crc kubenswrapper[4706]: I1206 14:16:25.521198 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9" path="/var/lib/kubelet/pods/352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9/volumes" Dec 06 14:16:25 crc kubenswrapper[4706]: I1206 14:16:25.522397 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="380fba54-3c03-4e8f-b702-6a2ad96c2744" path="/var/lib/kubelet/pods/380fba54-3c03-4e8f-b702-6a2ad96c2744/volumes" Dec 06 14:16:25 crc kubenswrapper[4706]: I1206 14:16:25.524618 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7490dd58-b444-46c5-ad33-b953af4e6617" path="/var/lib/kubelet/pods/7490dd58-b444-46c5-ad33-b953af4e6617/volumes" Dec 06 14:16:25 crc kubenswrapper[4706]: I1206 14:16:25.525870 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a86ee400-6a8a-4cc6-bdea-62a40954a8b2" path="/var/lib/kubelet/pods/a86ee400-6a8a-4cc6-bdea-62a40954a8b2/volumes" Dec 06 14:16:28 crc kubenswrapper[4706]: E1206 14:16:28.615218 4706 secret.go:188] Couldn't get secret openstack-operators/infra-operator-controller-manager-service-cert: secret "infra-operator-controller-manager-service-cert" not found Dec 06 14:16:28 crc kubenswrapper[4706]: E1206 14:16:28.615519 4706 secret.go:188] Couldn't get secret openstack-operators/infra-operator-controller-manager-service-cert: secret "infra-operator-controller-manager-service-cert" not found Dec 06 14:16:28 crc kubenswrapper[4706]: E1206 14:16:28.624509 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-webhook-cert podName:e430699f-156a-4c0c-b038-06a27b691852 nodeName:}" failed. No retries permitted until 2025-12-06 14:16:29.124472612 +0000 UTC m=+1231.960265479 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-webhook-cert") pod "infra-operator-controller-manager-d96b7b8b9-qpq4s" (UID: "e430699f-156a-4c0c-b038-06a27b691852") : secret "infra-operator-controller-manager-service-cert" not found Dec 06 14:16:28 crc kubenswrapper[4706]: E1206 14:16:28.624568 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-apiservice-cert podName:e430699f-156a-4c0c-b038-06a27b691852 nodeName:}" failed. No retries permitted until 2025-12-06 14:16:29.124551614 +0000 UTC m=+1231.960344471 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-apiservice-cert") pod "infra-operator-controller-manager-d96b7b8b9-qpq4s" (UID: "e430699f-156a-4c0c-b038-06a27b691852") : secret "infra-operator-controller-manager-service-cert" not found Dec 06 14:16:28 crc kubenswrapper[4706]: I1206 14:16:28.637375 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s"] Dec 06 14:16:28 crc kubenswrapper[4706]: I1206 14:16:28.773465 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" podUID="e430699f-156a-4c0c-b038-06a27b691852" containerName="manager" containerID="cri-o://1a4db1c512a94198a3f6a86efb734f507e64d2c61348277262b04f52e017216e" gracePeriod=10 Dec 06 14:16:28 crc kubenswrapper[4706]: I1206 14:16:28.930537 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-7vllp"] Dec 06 14:16:28 crc kubenswrapper[4706]: I1206 14:16:28.930753 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-7vllp" podUID="e5625341-f864-4891-87e4-833fbd15d8da" containerName="registry-server" containerID="cri-o://dbec3957b7f16b9db17d46328986557b6b4b0e235d7b57e7dd9eddc7080a3bc1" gracePeriod=30 Dec 06 14:16:28 crc kubenswrapper[4706]: I1206 14:16:28.985112 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d"] Dec 06 14:16:28 crc kubenswrapper[4706]: I1206 14:16:28.988903 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/7c8268737cab4bb480b8c6360540a07b7a71682e502fa0b9fb67d6a17a6kf5d"] Dec 06 14:16:29 crc kubenswrapper[4706]: E1206 14:16:29.222548 4706 secret.go:188] Couldn't get secret openstack-operators/infra-operator-controller-manager-service-cert: secret "infra-operator-controller-manager-service-cert" not found Dec 06 14:16:29 crc kubenswrapper[4706]: E1206 14:16:29.222604 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-webhook-cert podName:e430699f-156a-4c0c-b038-06a27b691852 nodeName:}" failed. No retries permitted until 2025-12-06 14:16:30.222588826 +0000 UTC m=+1233.058381683 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-webhook-cert") pod "infra-operator-controller-manager-d96b7b8b9-qpq4s" (UID: "e430699f-156a-4c0c-b038-06a27b691852") : secret "infra-operator-controller-manager-service-cert" not found Dec 06 14:16:29 crc kubenswrapper[4706]: E1206 14:16:29.222949 4706 secret.go:188] Couldn't get secret openstack-operators/infra-operator-controller-manager-service-cert: secret "infra-operator-controller-manager-service-cert" not found Dec 06 14:16:29 crc kubenswrapper[4706]: E1206 14:16:29.222975 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-apiservice-cert podName:e430699f-156a-4c0c-b038-06a27b691852 nodeName:}" failed. No retries permitted until 2025-12-06 14:16:30.222967036 +0000 UTC m=+1233.058759893 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-apiservice-cert") pod "infra-operator-controller-manager-d96b7b8b9-qpq4s" (UID: "e430699f-156a-4c0c-b038-06a27b691852") : secret "infra-operator-controller-manager-service-cert" not found Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.271907 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.343877 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-7vllp" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.424747 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zrn4\" (UniqueName: \"kubernetes.io/projected/e430699f-156a-4c0c-b038-06a27b691852-kube-api-access-7zrn4\") pod \"e430699f-156a-4c0c-b038-06a27b691852\" (UID: \"e430699f-156a-4c0c-b038-06a27b691852\") " Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.424797 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-webhook-cert\") pod \"e430699f-156a-4c0c-b038-06a27b691852\" (UID: \"e430699f-156a-4c0c-b038-06a27b691852\") " Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.424833 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-apiservice-cert\") pod \"e430699f-156a-4c0c-b038-06a27b691852\" (UID: \"e430699f-156a-4c0c-b038-06a27b691852\") " Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.424931 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2b6sc\" (UniqueName: \"kubernetes.io/projected/e5625341-f864-4891-87e4-833fbd15d8da-kube-api-access-2b6sc\") pod \"e5625341-f864-4891-87e4-833fbd15d8da\" (UID: \"e5625341-f864-4891-87e4-833fbd15d8da\") " Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.429822 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e430699f-156a-4c0c-b038-06a27b691852-kube-api-access-7zrn4" (OuterVolumeSpecName: "kube-api-access-7zrn4") pod "e430699f-156a-4c0c-b038-06a27b691852" (UID: "e430699f-156a-4c0c-b038-06a27b691852"). InnerVolumeSpecName "kube-api-access-7zrn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.429956 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5625341-f864-4891-87e4-833fbd15d8da-kube-api-access-2b6sc" (OuterVolumeSpecName: "kube-api-access-2b6sc") pod "e5625341-f864-4891-87e4-833fbd15d8da" (UID: "e5625341-f864-4891-87e4-833fbd15d8da"). InnerVolumeSpecName "kube-api-access-2b6sc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.430928 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "e430699f-156a-4c0c-b038-06a27b691852" (UID: "e430699f-156a-4c0c-b038-06a27b691852"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.431642 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "e430699f-156a-4c0c-b038-06a27b691852" (UID: "e430699f-156a-4c0c-b038-06a27b691852"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.526123 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2b6sc\" (UniqueName: \"kubernetes.io/projected/e5625341-f864-4891-87e4-833fbd15d8da-kube-api-access-2b6sc\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.526152 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zrn4\" (UniqueName: \"kubernetes.io/projected/e430699f-156a-4c0c-b038-06a27b691852-kube-api-access-7zrn4\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.526163 4706 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.526192 4706 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e430699f-156a-4c0c-b038-06a27b691852-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.528020 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44f63b09-00e7-45a5-b5c1-ffb282316231" path="/var/lib/kubelet/pods/44f63b09-00e7-45a5-b5c1-ffb282316231/volumes" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.779819 4706 generic.go:334] "Generic (PLEG): container finished" podID="e430699f-156a-4c0c-b038-06a27b691852" containerID="1a4db1c512a94198a3f6a86efb734f507e64d2c61348277262b04f52e017216e" exitCode=0 Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.779858 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" event={"ID":"e430699f-156a-4c0c-b038-06a27b691852","Type":"ContainerDied","Data":"1a4db1c512a94198a3f6a86efb734f507e64d2c61348277262b04f52e017216e"} Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.779866 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.780802 4706 scope.go:117] "RemoveContainer" containerID="1a4db1c512a94198a3f6a86efb734f507e64d2c61348277262b04f52e017216e" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.780769 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s" event={"ID":"e430699f-156a-4c0c-b038-06a27b691852","Type":"ContainerDied","Data":"8bbfdabd7738dc8cba307bf56a29ce24295c719e7979a26b4f7b395b237638ae"} Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.782331 4706 generic.go:334] "Generic (PLEG): container finished" podID="e5625341-f864-4891-87e4-833fbd15d8da" containerID="dbec3957b7f16b9db17d46328986557b6b4b0e235d7b57e7dd9eddc7080a3bc1" exitCode=0 Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.782360 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-7vllp" event={"ID":"e5625341-f864-4891-87e4-833fbd15d8da","Type":"ContainerDied","Data":"dbec3957b7f16b9db17d46328986557b6b4b0e235d7b57e7dd9eddc7080a3bc1"} Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.782382 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-7vllp" event={"ID":"e5625341-f864-4891-87e4-833fbd15d8da","Type":"ContainerDied","Data":"6b5347ee62e11dba51d6ff45845a48405c3fe7afd92a96bf451c561781b18a57"} Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.782393 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-7vllp" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.800865 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-7vllp"] Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.807170 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-7vllp"] Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.807375 4706 scope.go:117] "RemoveContainer" containerID="1a4db1c512a94198a3f6a86efb734f507e64d2c61348277262b04f52e017216e" Dec 06 14:16:29 crc kubenswrapper[4706]: E1206 14:16:29.807996 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a4db1c512a94198a3f6a86efb734f507e64d2c61348277262b04f52e017216e\": container with ID starting with 1a4db1c512a94198a3f6a86efb734f507e64d2c61348277262b04f52e017216e not found: ID does not exist" containerID="1a4db1c512a94198a3f6a86efb734f507e64d2c61348277262b04f52e017216e" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.808025 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a4db1c512a94198a3f6a86efb734f507e64d2c61348277262b04f52e017216e"} err="failed to get container status \"1a4db1c512a94198a3f6a86efb734f507e64d2c61348277262b04f52e017216e\": rpc error: code = NotFound desc = could not find container \"1a4db1c512a94198a3f6a86efb734f507e64d2c61348277262b04f52e017216e\": container with ID starting with 1a4db1c512a94198a3f6a86efb734f507e64d2c61348277262b04f52e017216e not found: ID does not exist" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.808044 4706 scope.go:117] "RemoveContainer" containerID="dbec3957b7f16b9db17d46328986557b6b4b0e235d7b57e7dd9eddc7080a3bc1" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.815629 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s"] Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.819722 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-controller-manager-d96b7b8b9-qpq4s"] Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.824439 4706 scope.go:117] "RemoveContainer" containerID="dbec3957b7f16b9db17d46328986557b6b4b0e235d7b57e7dd9eddc7080a3bc1" Dec 06 14:16:29 crc kubenswrapper[4706]: E1206 14:16:29.824833 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbec3957b7f16b9db17d46328986557b6b4b0e235d7b57e7dd9eddc7080a3bc1\": container with ID starting with dbec3957b7f16b9db17d46328986557b6b4b0e235d7b57e7dd9eddc7080a3bc1 not found: ID does not exist" containerID="dbec3957b7f16b9db17d46328986557b6b4b0e235d7b57e7dd9eddc7080a3bc1" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.824860 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbec3957b7f16b9db17d46328986557b6b4b0e235d7b57e7dd9eddc7080a3bc1"} err="failed to get container status \"dbec3957b7f16b9db17d46328986557b6b4b0e235d7b57e7dd9eddc7080a3bc1\": rpc error: code = NotFound desc = could not find container \"dbec3957b7f16b9db17d46328986557b6b4b0e235d7b57e7dd9eddc7080a3bc1\": container with ID starting with dbec3957b7f16b9db17d46328986557b6b4b0e235d7b57e7dd9eddc7080a3bc1 not found: ID does not exist" Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.994881 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw"] Dec 06 14:16:29 crc kubenswrapper[4706]: I1206 14:16:29.995090 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" podUID="a7081ae6-22a4-485d-93af-0662f73654f3" containerName="manager" containerID="cri-o://fe6eb7cc2cff3d138effb0cba8430cac91c2ea7aaaacb6fa5f74476dd0fb810f" gracePeriod=10 Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.327445 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-cq77l"] Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.327646 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-cq77l" podUID="ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9" containerName="registry-server" containerID="cri-o://e80df8ebfe4ceeadfdb9dd4e004f4ceba231697b614b8ea13afe092239419dd1" gracePeriod=30 Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.348910 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2"] Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.352082 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/e9e5855e5cd3b19ff946d1a783ccd6861442182df01f7778e40ab7fce77zfz2"] Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.486985 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.642150 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnwl6\" (UniqueName: \"kubernetes.io/projected/a7081ae6-22a4-485d-93af-0662f73654f3-kube-api-access-qnwl6\") pod \"a7081ae6-22a4-485d-93af-0662f73654f3\" (UID: \"a7081ae6-22a4-485d-93af-0662f73654f3\") " Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.642246 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7081ae6-22a4-485d-93af-0662f73654f3-apiservice-cert\") pod \"a7081ae6-22a4-485d-93af-0662f73654f3\" (UID: \"a7081ae6-22a4-485d-93af-0662f73654f3\") " Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.642339 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7081ae6-22a4-485d-93af-0662f73654f3-webhook-cert\") pod \"a7081ae6-22a4-485d-93af-0662f73654f3\" (UID: \"a7081ae6-22a4-485d-93af-0662f73654f3\") " Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.645988 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7081ae6-22a4-485d-93af-0662f73654f3-kube-api-access-qnwl6" (OuterVolumeSpecName: "kube-api-access-qnwl6") pod "a7081ae6-22a4-485d-93af-0662f73654f3" (UID: "a7081ae6-22a4-485d-93af-0662f73654f3"). InnerVolumeSpecName "kube-api-access-qnwl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.647555 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7081ae6-22a4-485d-93af-0662f73654f3-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "a7081ae6-22a4-485d-93af-0662f73654f3" (UID: "a7081ae6-22a4-485d-93af-0662f73654f3"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.649484 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7081ae6-22a4-485d-93af-0662f73654f3-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "a7081ae6-22a4-485d-93af-0662f73654f3" (UID: "a7081ae6-22a4-485d-93af-0662f73654f3"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.707747 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-cq77l" Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.744257 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnwl6\" (UniqueName: \"kubernetes.io/projected/a7081ae6-22a4-485d-93af-0662f73654f3-kube-api-access-qnwl6\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.744298 4706 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7081ae6-22a4-485d-93af-0662f73654f3-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.744366 4706 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7081ae6-22a4-485d-93af-0662f73654f3-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.795161 4706 generic.go:334] "Generic (PLEG): container finished" podID="ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9" containerID="e80df8ebfe4ceeadfdb9dd4e004f4ceba231697b614b8ea13afe092239419dd1" exitCode=0 Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.795276 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-cq77l" Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.795792 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-cq77l" event={"ID":"ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9","Type":"ContainerDied","Data":"e80df8ebfe4ceeadfdb9dd4e004f4ceba231697b614b8ea13afe092239419dd1"} Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.795838 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-cq77l" event={"ID":"ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9","Type":"ContainerDied","Data":"6b6cc2e086f2b7dd3dc846198136979fd5699d676f1330b19f03a840a927ab07"} Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.795859 4706 scope.go:117] "RemoveContainer" containerID="e80df8ebfe4ceeadfdb9dd4e004f4ceba231697b614b8ea13afe092239419dd1" Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.802885 4706 generic.go:334] "Generic (PLEG): container finished" podID="a7081ae6-22a4-485d-93af-0662f73654f3" containerID="fe6eb7cc2cff3d138effb0cba8430cac91c2ea7aaaacb6fa5f74476dd0fb810f" exitCode=0 Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.802926 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.802923 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" event={"ID":"a7081ae6-22a4-485d-93af-0662f73654f3","Type":"ContainerDied","Data":"fe6eb7cc2cff3d138effb0cba8430cac91c2ea7aaaacb6fa5f74476dd0fb810f"} Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.803050 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw" event={"ID":"a7081ae6-22a4-485d-93af-0662f73654f3","Type":"ContainerDied","Data":"187aa8aee4ab2b433c077849f82dfd6509bdd4d5ef605d5d48147562de2e9972"} Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.819332 4706 scope.go:117] "RemoveContainer" containerID="e80df8ebfe4ceeadfdb9dd4e004f4ceba231697b614b8ea13afe092239419dd1" Dec 06 14:16:30 crc kubenswrapper[4706]: E1206 14:16:30.819739 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e80df8ebfe4ceeadfdb9dd4e004f4ceba231697b614b8ea13afe092239419dd1\": container with ID starting with e80df8ebfe4ceeadfdb9dd4e004f4ceba231697b614b8ea13afe092239419dd1 not found: ID does not exist" containerID="e80df8ebfe4ceeadfdb9dd4e004f4ceba231697b614b8ea13afe092239419dd1" Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.819774 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e80df8ebfe4ceeadfdb9dd4e004f4ceba231697b614b8ea13afe092239419dd1"} err="failed to get container status \"e80df8ebfe4ceeadfdb9dd4e004f4ceba231697b614b8ea13afe092239419dd1\": rpc error: code = NotFound desc = could not find container \"e80df8ebfe4ceeadfdb9dd4e004f4ceba231697b614b8ea13afe092239419dd1\": container with ID starting with e80df8ebfe4ceeadfdb9dd4e004f4ceba231697b614b8ea13afe092239419dd1 not found: ID does not exist" Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.819796 4706 scope.go:117] "RemoveContainer" containerID="fe6eb7cc2cff3d138effb0cba8430cac91c2ea7aaaacb6fa5f74476dd0fb810f" Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.830720 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw"] Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.834866 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-759c788948-76bcw"] Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.840840 4706 scope.go:117] "RemoveContainer" containerID="fe6eb7cc2cff3d138effb0cba8430cac91c2ea7aaaacb6fa5f74476dd0fb810f" Dec 06 14:16:30 crc kubenswrapper[4706]: E1206 14:16:30.841493 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe6eb7cc2cff3d138effb0cba8430cac91c2ea7aaaacb6fa5f74476dd0fb810f\": container with ID starting with fe6eb7cc2cff3d138effb0cba8430cac91c2ea7aaaacb6fa5f74476dd0fb810f not found: ID does not exist" containerID="fe6eb7cc2cff3d138effb0cba8430cac91c2ea7aaaacb6fa5f74476dd0fb810f" Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.841543 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe6eb7cc2cff3d138effb0cba8430cac91c2ea7aaaacb6fa5f74476dd0fb810f"} err="failed to get container status \"fe6eb7cc2cff3d138effb0cba8430cac91c2ea7aaaacb6fa5f74476dd0fb810f\": rpc error: code = NotFound desc = could not find container \"fe6eb7cc2cff3d138effb0cba8430cac91c2ea7aaaacb6fa5f74476dd0fb810f\": container with ID starting with fe6eb7cc2cff3d138effb0cba8430cac91c2ea7aaaacb6fa5f74476dd0fb810f not found: ID does not exist" Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.845005 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2j9r\" (UniqueName: \"kubernetes.io/projected/ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9-kube-api-access-v2j9r\") pod \"ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9\" (UID: \"ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9\") " Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.849640 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9-kube-api-access-v2j9r" (OuterVolumeSpecName: "kube-api-access-v2j9r") pod "ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9" (UID: "ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9"). InnerVolumeSpecName "kube-api-access-v2j9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:16:30 crc kubenswrapper[4706]: I1206 14:16:30.946776 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2j9r\" (UniqueName: \"kubernetes.io/projected/ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9-kube-api-access-v2j9r\") on node \"crc\" DevicePath \"\"" Dec 06 14:16:31 crc kubenswrapper[4706]: I1206 14:16:31.141241 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-cq77l"] Dec 06 14:16:31 crc kubenswrapper[4706]: I1206 14:16:31.149792 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-cq77l"] Dec 06 14:16:31 crc kubenswrapper[4706]: I1206 14:16:31.522538 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e63f666-ace4-417b-ab05-50be328e8a64" path="/var/lib/kubelet/pods/3e63f666-ace4-417b-ab05-50be328e8a64/volumes" Dec 06 14:16:31 crc kubenswrapper[4706]: I1206 14:16:31.523947 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7081ae6-22a4-485d-93af-0662f73654f3" path="/var/lib/kubelet/pods/a7081ae6-22a4-485d-93af-0662f73654f3/volumes" Dec 06 14:16:31 crc kubenswrapper[4706]: I1206 14:16:31.525015 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e430699f-156a-4c0c-b038-06a27b691852" path="/var/lib/kubelet/pods/e430699f-156a-4c0c-b038-06a27b691852/volumes" Dec 06 14:16:31 crc kubenswrapper[4706]: I1206 14:16:31.526519 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5625341-f864-4891-87e4-833fbd15d8da" path="/var/lib/kubelet/pods/e5625341-f864-4891-87e4-833fbd15d8da/volumes" Dec 06 14:16:31 crc kubenswrapper[4706]: I1206 14:16:31.526908 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9" path="/var/lib/kubelet/pods/ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9/volumes" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.152428 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-b5wp7/must-gather-hlcts"] Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153201 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23f8c907-7a56-4da3-aefb-929a43497eb2" containerName="setup-container" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153217 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="23f8c907-7a56-4da3-aefb-929a43497eb2" containerName="setup-container" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153233 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a993fd-482d-4efd-938a-2014dda4aca1" containerName="memcached" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153241 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a993fd-482d-4efd-938a-2014dda4aca1" containerName="memcached" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153251 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18cc7358-0421-415c-a37b-6dba9a764d93" containerName="operator" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153260 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="18cc7358-0421-415c-a37b-6dba9a764d93" containerName="operator" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153271 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7490dd58-b444-46c5-ad33-b953af4e6617" containerName="registry-server" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153278 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="7490dd58-b444-46c5-ad33-b953af4e6617" containerName="registry-server" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153287 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5625341-f864-4891-87e4-833fbd15d8da" containerName="registry-server" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153294 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5625341-f864-4891-87e4-833fbd15d8da" containerName="registry-server" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153432 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9" containerName="registry-server" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153441 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9" containerName="registry-server" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153451 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="252d778a-eb1a-4519-9325-e79910544f1d" containerName="mysql-bootstrap" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153459 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="252d778a-eb1a-4519-9325-e79910544f1d" containerName="mysql-bootstrap" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153469 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a86ee400-6a8a-4cc6-bdea-62a40954a8b2" containerName="galera" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153477 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a86ee400-6a8a-4cc6-bdea-62a40954a8b2" containerName="galera" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153492 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40dc67dc-8180-452e-9345-a3e52b2adc15" containerName="manager" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153499 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="40dc67dc-8180-452e-9345-a3e52b2adc15" containerName="manager" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153512 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23f8c907-7a56-4da3-aefb-929a43497eb2" containerName="rabbitmq" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153519 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="23f8c907-7a56-4da3-aefb-929a43497eb2" containerName="rabbitmq" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153531 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ddad38d-0652-48ab-a70e-7eca376ba89d" containerName="keystone-api" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153538 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ddad38d-0652-48ab-a70e-7eca376ba89d" containerName="keystone-api" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153550 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5dce085-c674-49c0-883c-d1d788c1b70d" containerName="mariadb-account-delete" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153557 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5dce085-c674-49c0-883c-d1d788c1b70d" containerName="mariadb-account-delete" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153570 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7081ae6-22a4-485d-93af-0662f73654f3" containerName="manager" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153578 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7081ae6-22a4-485d-93af-0662f73654f3" containerName="manager" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153588 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b63157-bdec-47c2-a4b0-228f89541a2a" containerName="galera" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153596 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b63157-bdec-47c2-a4b0-228f89541a2a" containerName="galera" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153607 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5dce085-c674-49c0-883c-d1d788c1b70d" containerName="mariadb-account-delete" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153615 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5dce085-c674-49c0-883c-d1d788c1b70d" containerName="mariadb-account-delete" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153630 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e430699f-156a-4c0c-b038-06a27b691852" containerName="manager" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153638 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e430699f-156a-4c0c-b038-06a27b691852" containerName="manager" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153647 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="252d778a-eb1a-4519-9325-e79910544f1d" containerName="galera" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153654 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="252d778a-eb1a-4519-9325-e79910544f1d" containerName="galera" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153662 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c36a0ce1-c1a6-4529-9579-a48508a688fc" containerName="registry-server" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153670 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c36a0ce1-c1a6-4529-9579-a48508a688fc" containerName="registry-server" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153681 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b63157-bdec-47c2-a4b0-228f89541a2a" containerName="mysql-bootstrap" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153689 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b63157-bdec-47c2-a4b0-228f89541a2a" containerName="mysql-bootstrap" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153705 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a86ee400-6a8a-4cc6-bdea-62a40954a8b2" containerName="mysql-bootstrap" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153713 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a86ee400-6a8a-4cc6-bdea-62a40954a8b2" containerName="mysql-bootstrap" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153724 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c97961d-a743-4c7d-9bde-f319fea2af90" containerName="manager" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153734 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c97961d-a743-4c7d-9bde-f319fea2af90" containerName="manager" Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.153745 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9" containerName="registry-server" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153753 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9" containerName="registry-server" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153867 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="252d778a-eb1a-4519-9325-e79910544f1d" containerName="galera" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153884 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="352a19ee-27fc-4b0f-b7c7-e5eda8b88dc9" containerName="registry-server" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153896 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c97961d-a743-4c7d-9bde-f319fea2af90" containerName="manager" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153910 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3ea9d97-694e-43f8-8747-64c34bdab8dd" containerName="cinder-volume" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153919 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="7490dd58-b444-46c5-ad33-b953af4e6617" containerName="registry-server" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153928 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ddad38d-0652-48ab-a70e-7eca376ba89d" containerName="keystone-api" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153938 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e430699f-156a-4c0c-b038-06a27b691852" containerName="manager" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153946 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="41b63157-bdec-47c2-a4b0-228f89541a2a" containerName="galera" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153955 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5dce085-c674-49c0-883c-d1d788c1b70d" containerName="mariadb-account-delete" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153965 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef80af1b-73e0-45f0-bcd9-25b8cf2d17f9" containerName="registry-server" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153974 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5625341-f864-4891-87e4-833fbd15d8da" containerName="registry-server" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.153989 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="76a993fd-482d-4efd-938a-2014dda4aca1" containerName="memcached" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.154004 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="40dc67dc-8180-452e-9345-a3e52b2adc15" containerName="manager" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.154015 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5dce085-c674-49c0-883c-d1d788c1b70d" containerName="mariadb-account-delete" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.154026 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="18cc7358-0421-415c-a37b-6dba9a764d93" containerName="operator" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.154037 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c36a0ce1-c1a6-4529-9579-a48508a688fc" containerName="registry-server" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.154047 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7081ae6-22a4-485d-93af-0662f73654f3" containerName="manager" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.154058 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a86ee400-6a8a-4cc6-bdea-62a40954a8b2" containerName="galera" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.154069 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="23f8c907-7a56-4da3-aefb-929a43497eb2" containerName="rabbitmq" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.154739 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5wp7/must-gather-hlcts" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.159194 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-b5wp7"/"openshift-service-ca.crt" Dec 06 14:16:44 crc kubenswrapper[4706]: W1206 14:16:44.164096 4706 reflector.go:561] object-"openshift-must-gather-b5wp7"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-must-gather-b5wp7": no relationship found between node 'crc' and this object Dec 06 14:16:44 crc kubenswrapper[4706]: E1206 14:16:44.164173 4706 reflector.go:158] "Unhandled Error" err="object-\"openshift-must-gather-b5wp7\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-must-gather-b5wp7\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.185414 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-b5wp7/must-gather-hlcts"] Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.246410 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj9xk\" (UniqueName: \"kubernetes.io/projected/b5fa5197-a4ea-40c0-a8b3-d0c37f055f87-kube-api-access-dj9xk\") pod \"must-gather-hlcts\" (UID: \"b5fa5197-a4ea-40c0-a8b3-d0c37f055f87\") " pod="openshift-must-gather-b5wp7/must-gather-hlcts" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.246493 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b5fa5197-a4ea-40c0-a8b3-d0c37f055f87-must-gather-output\") pod \"must-gather-hlcts\" (UID: \"b5fa5197-a4ea-40c0-a8b3-d0c37f055f87\") " pod="openshift-must-gather-b5wp7/must-gather-hlcts" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.348109 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj9xk\" (UniqueName: \"kubernetes.io/projected/b5fa5197-a4ea-40c0-a8b3-d0c37f055f87-kube-api-access-dj9xk\") pod \"must-gather-hlcts\" (UID: \"b5fa5197-a4ea-40c0-a8b3-d0c37f055f87\") " pod="openshift-must-gather-b5wp7/must-gather-hlcts" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.348198 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b5fa5197-a4ea-40c0-a8b3-d0c37f055f87-must-gather-output\") pod \"must-gather-hlcts\" (UID: \"b5fa5197-a4ea-40c0-a8b3-d0c37f055f87\") " pod="openshift-must-gather-b5wp7/must-gather-hlcts" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.348770 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b5fa5197-a4ea-40c0-a8b3-d0c37f055f87-must-gather-output\") pod \"must-gather-hlcts\" (UID: \"b5fa5197-a4ea-40c0-a8b3-d0c37f055f87\") " pod="openshift-must-gather-b5wp7/must-gather-hlcts" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.973579 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-b5wp7"/"kube-root-ca.crt" Dec 06 14:16:44 crc kubenswrapper[4706]: I1206 14:16:44.991469 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj9xk\" (UniqueName: \"kubernetes.io/projected/b5fa5197-a4ea-40c0-a8b3-d0c37f055f87-kube-api-access-dj9xk\") pod \"must-gather-hlcts\" (UID: \"b5fa5197-a4ea-40c0-a8b3-d0c37f055f87\") " pod="openshift-must-gather-b5wp7/must-gather-hlcts" Dec 06 14:16:45 crc kubenswrapper[4706]: I1206 14:16:45.078863 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5wp7/must-gather-hlcts" Dec 06 14:16:45 crc kubenswrapper[4706]: I1206 14:16:45.371345 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-b5wp7/must-gather-hlcts"] Dec 06 14:16:45 crc kubenswrapper[4706]: I1206 14:16:45.935672 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5wp7/must-gather-hlcts" event={"ID":"b5fa5197-a4ea-40c0-a8b3-d0c37f055f87","Type":"ContainerStarted","Data":"87c5e408d5af9a55014c9d8c3d95c1c7103b5a2886561a0d8b4383b97c542ea7"} Dec 06 14:16:49 crc kubenswrapper[4706]: I1206 14:16:49.981846 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5wp7/must-gather-hlcts" event={"ID":"b5fa5197-a4ea-40c0-a8b3-d0c37f055f87","Type":"ContainerStarted","Data":"5509461c44b25ec7fba6c763d1783f79fe8f60a1e638481a6d84ca197974274c"} Dec 06 14:16:49 crc kubenswrapper[4706]: I1206 14:16:49.982485 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5wp7/must-gather-hlcts" event={"ID":"b5fa5197-a4ea-40c0-a8b3-d0c37f055f87","Type":"ContainerStarted","Data":"5b0e9e52e710c46c43ce9c67aafd30cb18a6c26ced09352a68a2f97ccea6c11a"} Dec 06 14:16:50 crc kubenswrapper[4706]: I1206 14:16:50.004720 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-b5wp7/must-gather-hlcts" podStartSLOduration=2.312796523 podStartE2EDuration="6.00469626s" podCreationTimestamp="2025-12-06 14:16:44 +0000 UTC" firstStartedPulling="2025-12-06 14:16:45.38665629 +0000 UTC m=+1248.222449187" lastFinishedPulling="2025-12-06 14:16:49.078556027 +0000 UTC m=+1251.914348924" observedRunningTime="2025-12-06 14:16:50.000389856 +0000 UTC m=+1252.836182723" watchObservedRunningTime="2025-12-06 14:16:50.00469626 +0000 UTC m=+1252.840489157" Dec 06 14:16:58 crc kubenswrapper[4706]: I1206 14:16:58.688712 4706 scope.go:117] "RemoveContainer" containerID="55e1d3d6f27102508e4473f17e2112843cd9400f821dcce0566586254bcec43b" Dec 06 14:16:58 crc kubenswrapper[4706]: I1206 14:16:58.709533 4706 scope.go:117] "RemoveContainer" containerID="c911be28060746c5240912071547f69459b978a79557e8ec1d05b255d87fd594" Dec 06 14:16:58 crc kubenswrapper[4706]: I1206 14:16:58.734948 4706 scope.go:117] "RemoveContainer" containerID="3aa34f564128962b20c74ee2bbadd559a82b9b733956c2fee62daced7e00c835" Dec 06 14:16:58 crc kubenswrapper[4706]: I1206 14:16:58.748553 4706 scope.go:117] "RemoveContainer" containerID="b3ce0d041bb17a241d4703ada9ad4f49f6e0f6ac810b3f5d933da58a9e0236ec" Dec 06 14:16:58 crc kubenswrapper[4706]: I1206 14:16:58.773198 4706 scope.go:117] "RemoveContainer" containerID="7b784979a0eb31f9254a036717a835a82031cd97863423e2a7353a3578656358" Dec 06 14:16:58 crc kubenswrapper[4706]: I1206 14:16:58.795898 4706 scope.go:117] "RemoveContainer" containerID="84be07ce6389a4820d53a56b02ac8426f433e23df7e3b1b177c0cd30560e5e81" Dec 06 14:17:36 crc kubenswrapper[4706]: I1206 14:17:36.907895 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-p85f8_77b4e41c-830d-48c2-8d00-97b33c7e8a29/control-plane-machine-set-operator/0.log" Dec 06 14:17:36 crc kubenswrapper[4706]: I1206 14:17:36.960009 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9hxp8_4c02a475-2aec-4c9c-ba4d-f20059c09682/kube-rbac-proxy/0.log" Dec 06 14:17:37 crc kubenswrapper[4706]: I1206 14:17:37.036018 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9hxp8_4c02a475-2aec-4c9c-ba4d-f20059c09682/machine-api-operator/0.log" Dec 06 14:17:53 crc kubenswrapper[4706]: I1206 14:17:53.518846 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-bcgwr_d6c3a1b0-6512-40a8-aac5-bc6629793ab8/kube-rbac-proxy/0.log" Dec 06 14:17:53 crc kubenswrapper[4706]: I1206 14:17:53.606946 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-bcgwr_d6c3a1b0-6512-40a8-aac5-bc6629793ab8/controller/0.log" Dec 06 14:17:53 crc kubenswrapper[4706]: I1206 14:17:53.723106 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-5vmhj_7cd749c9-c857-4019-9fb3-f2c53f0b9be8/frr-k8s-webhook-server/0.log" Dec 06 14:17:53 crc kubenswrapper[4706]: I1206 14:17:53.803142 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-frr-files/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.003294 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-frr-files/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.003365 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-reloader/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.008072 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-metrics/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.041906 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-reloader/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.154167 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-frr-files/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.181730 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-reloader/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.197934 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-metrics/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.227178 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-metrics/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.374931 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-frr-files/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.377937 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-metrics/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.382456 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-reloader/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.411541 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/controller/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.542561 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/frr-metrics/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.542740 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/kube-rbac-proxy/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.635136 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/kube-rbac-proxy-frr/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.733838 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/reloader/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.829760 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-548bff8fd9-mt2r9_5acf3c41-82f9-4b93-8742-1b4975610426/manager/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.867507 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/frr/0.log" Dec 06 14:17:54 crc kubenswrapper[4706]: I1206 14:17:54.932861 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-56bc6d5c44-tg7hm_d16f9b04-736c-4331-8ad8-58d14bd858e6/webhook-server/0.log" Dec 06 14:17:55 crc kubenswrapper[4706]: I1206 14:17:55.075170 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-w6jgb_afd1dfce-966c-4506-bf87-3e2730857136/kube-rbac-proxy/0.log" Dec 06 14:17:55 crc kubenswrapper[4706]: I1206 14:17:55.447069 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-w6jgb_afd1dfce-966c-4506-bf87-3e2730857136/speaker/0.log" Dec 06 14:17:59 crc kubenswrapper[4706]: I1206 14:17:59.093400 4706 scope.go:117] "RemoveContainer" containerID="3f4a4f41c411bf999dc1d8b1b7a7ba8c746d7656f80bf16ad27f247a7153ab68" Dec 06 14:17:59 crc kubenswrapper[4706]: I1206 14:17:59.124943 4706 scope.go:117] "RemoveContainer" containerID="fba5757c3407cbacde9f45f2a39ffdd2c94fb9a6da8a4a223483c07c14dbd10e" Dec 06 14:17:59 crc kubenswrapper[4706]: I1206 14:17:59.146108 4706 scope.go:117] "RemoveContainer" containerID="e76355becb4e03558d1bb2f1a40134e7331e1c0f23b58bd50a00061f90f0f64f" Dec 06 14:17:59 crc kubenswrapper[4706]: I1206 14:17:59.170506 4706 scope.go:117] "RemoveContainer" containerID="445c087618e3f78c845de5174d8f25bd43aad6fe8f8d0210c0d8170406bdfd4b" Dec 06 14:17:59 crc kubenswrapper[4706]: I1206 14:17:59.194551 4706 scope.go:117] "RemoveContainer" containerID="74029b0dde685b98feafe6433cf565edfcc2b8cc0fb755106524a513baa7d500" Dec 06 14:17:59 crc kubenswrapper[4706]: I1206 14:17:59.211003 4706 scope.go:117] "RemoveContainer" containerID="82de2c90b4b986b488fcde85a41a0143e73cee5e2d671a6fa9a43b3c546bb84d" Dec 06 14:18:14 crc kubenswrapper[4706]: I1206 14:18:14.580346 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:18:14 crc kubenswrapper[4706]: I1206 14:18:14.580925 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:18:21 crc kubenswrapper[4706]: I1206 14:18:21.323052 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4/util/0.log" Dec 06 14:18:21 crc kubenswrapper[4706]: I1206 14:18:21.472351 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4/util/0.log" Dec 06 14:18:21 crc kubenswrapper[4706]: I1206 14:18:21.495823 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4/pull/0.log" Dec 06 14:18:21 crc kubenswrapper[4706]: I1206 14:18:21.503232 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4/pull/0.log" Dec 06 14:18:21 crc kubenswrapper[4706]: I1206 14:18:21.668800 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4/util/0.log" Dec 06 14:18:21 crc kubenswrapper[4706]: I1206 14:18:21.680183 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4/extract/0.log" Dec 06 14:18:21 crc kubenswrapper[4706]: I1206 14:18:21.686133 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4/pull/0.log" Dec 06 14:18:21 crc kubenswrapper[4706]: I1206 14:18:21.822962 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ssx77_6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110/extract-utilities/0.log" Dec 06 14:18:22 crc kubenswrapper[4706]: I1206 14:18:22.004773 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ssx77_6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110/extract-content/0.log" Dec 06 14:18:22 crc kubenswrapper[4706]: I1206 14:18:22.014539 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ssx77_6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110/extract-utilities/0.log" Dec 06 14:18:22 crc kubenswrapper[4706]: I1206 14:18:22.032271 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ssx77_6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110/extract-content/0.log" Dec 06 14:18:22 crc kubenswrapper[4706]: I1206 14:18:22.192848 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ssx77_6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110/extract-utilities/0.log" Dec 06 14:18:22 crc kubenswrapper[4706]: I1206 14:18:22.193039 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ssx77_6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110/extract-content/0.log" Dec 06 14:18:22 crc kubenswrapper[4706]: I1206 14:18:22.394287 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ggc6c_60caaf13-e871-4029-8614-dc43c10ac358/extract-utilities/0.log" Dec 06 14:18:22 crc kubenswrapper[4706]: I1206 14:18:22.468666 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ssx77_6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110/registry-server/0.log" Dec 06 14:18:22 crc kubenswrapper[4706]: I1206 14:18:22.520840 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ggc6c_60caaf13-e871-4029-8614-dc43c10ac358/extract-content/0.log" Dec 06 14:18:22 crc kubenswrapper[4706]: I1206 14:18:22.561713 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ggc6c_60caaf13-e871-4029-8614-dc43c10ac358/extract-utilities/0.log" Dec 06 14:18:22 crc kubenswrapper[4706]: I1206 14:18:22.567036 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ggc6c_60caaf13-e871-4029-8614-dc43c10ac358/extract-content/0.log" Dec 06 14:18:22 crc kubenswrapper[4706]: I1206 14:18:22.733553 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ggc6c_60caaf13-e871-4029-8614-dc43c10ac358/extract-utilities/0.log" Dec 06 14:18:22 crc kubenswrapper[4706]: I1206 14:18:22.739811 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ggc6c_60caaf13-e871-4029-8614-dc43c10ac358/extract-content/0.log" Dec 06 14:18:22 crc kubenswrapper[4706]: I1206 14:18:22.917049 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-sgfd9_92920365-098a-442a-87de-91aa1479ff1d/marketplace-operator/0.log" Dec 06 14:18:23 crc kubenswrapper[4706]: I1206 14:18:23.040080 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-62dkd_0514bb3a-f280-4208-9dff-ca5549dff0ca/extract-utilities/0.log" Dec 06 14:18:23 crc kubenswrapper[4706]: I1206 14:18:23.042615 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ggc6c_60caaf13-e871-4029-8614-dc43c10ac358/registry-server/0.log" Dec 06 14:18:23 crc kubenswrapper[4706]: I1206 14:18:23.159167 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-62dkd_0514bb3a-f280-4208-9dff-ca5549dff0ca/extract-content/0.log" Dec 06 14:18:23 crc kubenswrapper[4706]: I1206 14:18:23.186251 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-62dkd_0514bb3a-f280-4208-9dff-ca5549dff0ca/extract-utilities/0.log" Dec 06 14:18:23 crc kubenswrapper[4706]: I1206 14:18:23.200424 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-62dkd_0514bb3a-f280-4208-9dff-ca5549dff0ca/extract-content/0.log" Dec 06 14:18:23 crc kubenswrapper[4706]: I1206 14:18:23.384919 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-62dkd_0514bb3a-f280-4208-9dff-ca5549dff0ca/extract-utilities/0.log" Dec 06 14:18:23 crc kubenswrapper[4706]: I1206 14:18:23.389040 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-62dkd_0514bb3a-f280-4208-9dff-ca5549dff0ca/extract-content/0.log" Dec 06 14:18:23 crc kubenswrapper[4706]: I1206 14:18:23.399937 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-62dkd_0514bb3a-f280-4208-9dff-ca5549dff0ca/registry-server/0.log" Dec 06 14:18:23 crc kubenswrapper[4706]: I1206 14:18:23.552384 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hl9mm_8278b19e-f7eb-43ee-8d01-473f09b98a95/extract-utilities/0.log" Dec 06 14:18:23 crc kubenswrapper[4706]: I1206 14:18:23.746664 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hl9mm_8278b19e-f7eb-43ee-8d01-473f09b98a95/extract-utilities/0.log" Dec 06 14:18:23 crc kubenswrapper[4706]: I1206 14:18:23.753379 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hl9mm_8278b19e-f7eb-43ee-8d01-473f09b98a95/extract-content/0.log" Dec 06 14:18:23 crc kubenswrapper[4706]: I1206 14:18:23.804437 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hl9mm_8278b19e-f7eb-43ee-8d01-473f09b98a95/extract-content/0.log" Dec 06 14:18:23 crc kubenswrapper[4706]: I1206 14:18:23.921332 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hl9mm_8278b19e-f7eb-43ee-8d01-473f09b98a95/extract-utilities/0.log" Dec 06 14:18:23 crc kubenswrapper[4706]: I1206 14:18:23.947164 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hl9mm_8278b19e-f7eb-43ee-8d01-473f09b98a95/extract-content/0.log" Dec 06 14:18:24 crc kubenswrapper[4706]: I1206 14:18:24.233212 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hl9mm_8278b19e-f7eb-43ee-8d01-473f09b98a95/registry-server/0.log" Dec 06 14:18:44 crc kubenswrapper[4706]: I1206 14:18:44.580152 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:18:44 crc kubenswrapper[4706]: I1206 14:18:44.580899 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:18:59 crc kubenswrapper[4706]: I1206 14:18:59.255049 4706 scope.go:117] "RemoveContainer" containerID="e856fc6f0fb78df070ed34276cb3d2465ed9db0bea3b53083e26354b90448d7e" Dec 06 14:18:59 crc kubenswrapper[4706]: I1206 14:18:59.296790 4706 scope.go:117] "RemoveContainer" containerID="f9aa6c96348f40276b585ba07998ba23d7f79cde69151709d1958250cb56696b" Dec 06 14:18:59 crc kubenswrapper[4706]: I1206 14:18:59.349532 4706 scope.go:117] "RemoveContainer" containerID="5ae103d7a9afc14c67fbafd0cc97a7632ce9dfb1edc608eeb7e480ff5f9e3e06" Dec 06 14:18:59 crc kubenswrapper[4706]: I1206 14:18:59.367712 4706 scope.go:117] "RemoveContainer" containerID="5004a92cee42e60bd649ea2769194287aca089e4702405bd3cc825bcc85d5eec" Dec 06 14:18:59 crc kubenswrapper[4706]: I1206 14:18:59.386907 4706 scope.go:117] "RemoveContainer" containerID="e63b9aa8c618d24417a97f3b26af0df3138dd89c00e4ea8971234646322f6402" Dec 06 14:18:59 crc kubenswrapper[4706]: I1206 14:18:59.423548 4706 scope.go:117] "RemoveContainer" containerID="4dbdd10af7c3a934f9c28fb5839b4d5dde743af583343c6d9a9bf4ba03475274" Dec 06 14:18:59 crc kubenswrapper[4706]: I1206 14:18:59.454098 4706 scope.go:117] "RemoveContainer" containerID="f05fd2d6df14194e82d8bd3794000b922adb1c66d3841ffc646e63f50fe9eccb" Dec 06 14:19:14 crc kubenswrapper[4706]: I1206 14:19:14.580276 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:19:14 crc kubenswrapper[4706]: I1206 14:19:14.580936 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:19:14 crc kubenswrapper[4706]: I1206 14:19:14.580998 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 14:19:14 crc kubenswrapper[4706]: I1206 14:19:14.581790 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"541fe7945667e62148e85f3104c224b6e7d1a0ff6a7b9516e33ef2027678744f"} pod="openshift-machine-config-operator/machine-config-daemon-67v56" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 14:19:14 crc kubenswrapper[4706]: I1206 14:19:14.581875 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" containerID="cri-o://541fe7945667e62148e85f3104c224b6e7d1a0ff6a7b9516e33ef2027678744f" gracePeriod=600 Dec 06 14:19:15 crc kubenswrapper[4706]: I1206 14:19:15.072245 4706 generic.go:334] "Generic (PLEG): container finished" podID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerID="541fe7945667e62148e85f3104c224b6e7d1a0ff6a7b9516e33ef2027678744f" exitCode=0 Dec 06 14:19:15 crc kubenswrapper[4706]: I1206 14:19:15.072473 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerDied","Data":"541fe7945667e62148e85f3104c224b6e7d1a0ff6a7b9516e33ef2027678744f"} Dec 06 14:19:15 crc kubenswrapper[4706]: I1206 14:19:15.072891 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerStarted","Data":"2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f"} Dec 06 14:19:15 crc kubenswrapper[4706]: I1206 14:19:15.072938 4706 scope.go:117] "RemoveContainer" containerID="953455e6657a15cff411732f38c90183866bd015ab24e23297c405658c8efab6" Dec 06 14:19:27 crc kubenswrapper[4706]: E1206 14:19:27.742284 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5fa5197_a4ea_40c0_a8b3_d0c37f055f87.slice/crio-5b0e9e52e710c46c43ce9c67aafd30cb18a6c26ced09352a68a2f97ccea6c11a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5fa5197_a4ea_40c0_a8b3_d0c37f055f87.slice/crio-conmon-5b0e9e52e710c46c43ce9c67aafd30cb18a6c26ced09352a68a2f97ccea6c11a.scope\": RecentStats: unable to find data in memory cache]" Dec 06 14:19:28 crc kubenswrapper[4706]: I1206 14:19:28.179158 4706 generic.go:334] "Generic (PLEG): container finished" podID="b5fa5197-a4ea-40c0-a8b3-d0c37f055f87" containerID="5b0e9e52e710c46c43ce9c67aafd30cb18a6c26ced09352a68a2f97ccea6c11a" exitCode=0 Dec 06 14:19:28 crc kubenswrapper[4706]: I1206 14:19:28.179241 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5wp7/must-gather-hlcts" event={"ID":"b5fa5197-a4ea-40c0-a8b3-d0c37f055f87","Type":"ContainerDied","Data":"5b0e9e52e710c46c43ce9c67aafd30cb18a6c26ced09352a68a2f97ccea6c11a"} Dec 06 14:19:28 crc kubenswrapper[4706]: I1206 14:19:28.181817 4706 scope.go:117] "RemoveContainer" containerID="5b0e9e52e710c46c43ce9c67aafd30cb18a6c26ced09352a68a2f97ccea6c11a" Dec 06 14:19:28 crc kubenswrapper[4706]: I1206 14:19:28.684180 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b5wp7_must-gather-hlcts_b5fa5197-a4ea-40c0-a8b3-d0c37f055f87/gather/0.log" Dec 06 14:19:35 crc kubenswrapper[4706]: I1206 14:19:35.699838 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-b5wp7/must-gather-hlcts"] Dec 06 14:19:35 crc kubenswrapper[4706]: I1206 14:19:35.700636 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-b5wp7/must-gather-hlcts" podUID="b5fa5197-a4ea-40c0-a8b3-d0c37f055f87" containerName="copy" containerID="cri-o://5509461c44b25ec7fba6c763d1783f79fe8f60a1e638481a6d84ca197974274c" gracePeriod=2 Dec 06 14:19:35 crc kubenswrapper[4706]: I1206 14:19:35.703710 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-b5wp7/must-gather-hlcts"] Dec 06 14:19:36 crc kubenswrapper[4706]: I1206 14:19:36.030975 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b5wp7_must-gather-hlcts_b5fa5197-a4ea-40c0-a8b3-d0c37f055f87/copy/0.log" Dec 06 14:19:36 crc kubenswrapper[4706]: I1206 14:19:36.031684 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5wp7/must-gather-hlcts" Dec 06 14:19:36 crc kubenswrapper[4706]: I1206 14:19:36.177705 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj9xk\" (UniqueName: \"kubernetes.io/projected/b5fa5197-a4ea-40c0-a8b3-d0c37f055f87-kube-api-access-dj9xk\") pod \"b5fa5197-a4ea-40c0-a8b3-d0c37f055f87\" (UID: \"b5fa5197-a4ea-40c0-a8b3-d0c37f055f87\") " Dec 06 14:19:36 crc kubenswrapper[4706]: I1206 14:19:36.177781 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b5fa5197-a4ea-40c0-a8b3-d0c37f055f87-must-gather-output\") pod \"b5fa5197-a4ea-40c0-a8b3-d0c37f055f87\" (UID: \"b5fa5197-a4ea-40c0-a8b3-d0c37f055f87\") " Dec 06 14:19:36 crc kubenswrapper[4706]: I1206 14:19:36.191124 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5fa5197-a4ea-40c0-a8b3-d0c37f055f87-kube-api-access-dj9xk" (OuterVolumeSpecName: "kube-api-access-dj9xk") pod "b5fa5197-a4ea-40c0-a8b3-d0c37f055f87" (UID: "b5fa5197-a4ea-40c0-a8b3-d0c37f055f87"). InnerVolumeSpecName "kube-api-access-dj9xk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:19:36 crc kubenswrapper[4706]: I1206 14:19:36.235113 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b5wp7_must-gather-hlcts_b5fa5197-a4ea-40c0-a8b3-d0c37f055f87/copy/0.log" Dec 06 14:19:36 crc kubenswrapper[4706]: I1206 14:19:36.235537 4706 generic.go:334] "Generic (PLEG): container finished" podID="b5fa5197-a4ea-40c0-a8b3-d0c37f055f87" containerID="5509461c44b25ec7fba6c763d1783f79fe8f60a1e638481a6d84ca197974274c" exitCode=143 Dec 06 14:19:36 crc kubenswrapper[4706]: I1206 14:19:36.235578 4706 scope.go:117] "RemoveContainer" containerID="5509461c44b25ec7fba6c763d1783f79fe8f60a1e638481a6d84ca197974274c" Dec 06 14:19:36 crc kubenswrapper[4706]: I1206 14:19:36.235995 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5wp7/must-gather-hlcts" Dec 06 14:19:36 crc kubenswrapper[4706]: I1206 14:19:36.236788 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5fa5197-a4ea-40c0-a8b3-d0c37f055f87-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b5fa5197-a4ea-40c0-a8b3-d0c37f055f87" (UID: "b5fa5197-a4ea-40c0-a8b3-d0c37f055f87"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:19:36 crc kubenswrapper[4706]: I1206 14:19:36.250796 4706 scope.go:117] "RemoveContainer" containerID="5b0e9e52e710c46c43ce9c67aafd30cb18a6c26ced09352a68a2f97ccea6c11a" Dec 06 14:19:36 crc kubenswrapper[4706]: I1206 14:19:36.279870 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj9xk\" (UniqueName: \"kubernetes.io/projected/b5fa5197-a4ea-40c0-a8b3-d0c37f055f87-kube-api-access-dj9xk\") on node \"crc\" DevicePath \"\"" Dec 06 14:19:36 crc kubenswrapper[4706]: I1206 14:19:36.279903 4706 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b5fa5197-a4ea-40c0-a8b3-d0c37f055f87-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 06 14:19:36 crc kubenswrapper[4706]: I1206 14:19:36.286351 4706 scope.go:117] "RemoveContainer" containerID="5509461c44b25ec7fba6c763d1783f79fe8f60a1e638481a6d84ca197974274c" Dec 06 14:19:36 crc kubenswrapper[4706]: E1206 14:19:36.286864 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5509461c44b25ec7fba6c763d1783f79fe8f60a1e638481a6d84ca197974274c\": container with ID starting with 5509461c44b25ec7fba6c763d1783f79fe8f60a1e638481a6d84ca197974274c not found: ID does not exist" containerID="5509461c44b25ec7fba6c763d1783f79fe8f60a1e638481a6d84ca197974274c" Dec 06 14:19:36 crc kubenswrapper[4706]: I1206 14:19:36.286899 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5509461c44b25ec7fba6c763d1783f79fe8f60a1e638481a6d84ca197974274c"} err="failed to get container status \"5509461c44b25ec7fba6c763d1783f79fe8f60a1e638481a6d84ca197974274c\": rpc error: code = NotFound desc = could not find container \"5509461c44b25ec7fba6c763d1783f79fe8f60a1e638481a6d84ca197974274c\": container with ID starting with 5509461c44b25ec7fba6c763d1783f79fe8f60a1e638481a6d84ca197974274c not found: ID does not exist" Dec 06 14:19:36 crc kubenswrapper[4706]: I1206 14:19:36.286925 4706 scope.go:117] "RemoveContainer" containerID="5b0e9e52e710c46c43ce9c67aafd30cb18a6c26ced09352a68a2f97ccea6c11a" Dec 06 14:19:36 crc kubenswrapper[4706]: E1206 14:19:36.287410 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b0e9e52e710c46c43ce9c67aafd30cb18a6c26ced09352a68a2f97ccea6c11a\": container with ID starting with 5b0e9e52e710c46c43ce9c67aafd30cb18a6c26ced09352a68a2f97ccea6c11a not found: ID does not exist" containerID="5b0e9e52e710c46c43ce9c67aafd30cb18a6c26ced09352a68a2f97ccea6c11a" Dec 06 14:19:36 crc kubenswrapper[4706]: I1206 14:19:36.287440 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b0e9e52e710c46c43ce9c67aafd30cb18a6c26ced09352a68a2f97ccea6c11a"} err="failed to get container status \"5b0e9e52e710c46c43ce9c67aafd30cb18a6c26ced09352a68a2f97ccea6c11a\": rpc error: code = NotFound desc = could not find container \"5b0e9e52e710c46c43ce9c67aafd30cb18a6c26ced09352a68a2f97ccea6c11a\": container with ID starting with 5b0e9e52e710c46c43ce9c67aafd30cb18a6c26ced09352a68a2f97ccea6c11a not found: ID does not exist" Dec 06 14:19:37 crc kubenswrapper[4706]: I1206 14:19:37.524557 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5fa5197-a4ea-40c0-a8b3-d0c37f055f87" path="/var/lib/kubelet/pods/b5fa5197-a4ea-40c0-a8b3-d0c37f055f87/volumes" Dec 06 14:19:59 crc kubenswrapper[4706]: I1206 14:19:59.543424 4706 scope.go:117] "RemoveContainer" containerID="5e8086afc9353a5159ca0256b01989180a096754f120731aa1e5716bc8a0eec6" Dec 06 14:19:59 crc kubenswrapper[4706]: I1206 14:19:59.617490 4706 scope.go:117] "RemoveContainer" containerID="06e0cc770ece8cab7566a1e52eda59fad9955394bd977916b8b619b901688275" Dec 06 14:19:59 crc kubenswrapper[4706]: I1206 14:19:59.648745 4706 scope.go:117] "RemoveContainer" containerID="b3d2693f98db9ba176c3d7acbd4cb1039e3723d4e5843e88fe8edf7af40c5015" Dec 06 14:21:14 crc kubenswrapper[4706]: I1206 14:21:14.580631 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:21:14 crc kubenswrapper[4706]: I1206 14:21:14.582823 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:21:44 crc kubenswrapper[4706]: I1206 14:21:44.580196 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:21:44 crc kubenswrapper[4706]: I1206 14:21:44.580733 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.396973 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vs56v"] Dec 06 14:21:59 crc kubenswrapper[4706]: E1206 14:21:59.397482 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5fa5197-a4ea-40c0-a8b3-d0c37f055f87" containerName="copy" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.397497 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5fa5197-a4ea-40c0-a8b3-d0c37f055f87" containerName="copy" Dec 06 14:21:59 crc kubenswrapper[4706]: E1206 14:21:59.397512 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5fa5197-a4ea-40c0-a8b3-d0c37f055f87" containerName="gather" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.397520 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5fa5197-a4ea-40c0-a8b3-d0c37f055f87" containerName="gather" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.397664 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5fa5197-a4ea-40c0-a8b3-d0c37f055f87" containerName="gather" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.397677 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5fa5197-a4ea-40c0-a8b3-d0c37f055f87" containerName="copy" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.398514 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vs56v" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.422444 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vs56v"] Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.453861 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a92752a-7051-4fe7-bba9-f106dd557e59-catalog-content\") pod \"redhat-operators-vs56v\" (UID: \"7a92752a-7051-4fe7-bba9-f106dd557e59\") " pod="openshift-marketplace/redhat-operators-vs56v" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.453922 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5dg8\" (UniqueName: \"kubernetes.io/projected/7a92752a-7051-4fe7-bba9-f106dd557e59-kube-api-access-n5dg8\") pod \"redhat-operators-vs56v\" (UID: \"7a92752a-7051-4fe7-bba9-f106dd557e59\") " pod="openshift-marketplace/redhat-operators-vs56v" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.454063 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a92752a-7051-4fe7-bba9-f106dd557e59-utilities\") pod \"redhat-operators-vs56v\" (UID: \"7a92752a-7051-4fe7-bba9-f106dd557e59\") " pod="openshift-marketplace/redhat-operators-vs56v" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.555852 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a92752a-7051-4fe7-bba9-f106dd557e59-catalog-content\") pod \"redhat-operators-vs56v\" (UID: \"7a92752a-7051-4fe7-bba9-f106dd557e59\") " pod="openshift-marketplace/redhat-operators-vs56v" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.555911 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5dg8\" (UniqueName: \"kubernetes.io/projected/7a92752a-7051-4fe7-bba9-f106dd557e59-kube-api-access-n5dg8\") pod \"redhat-operators-vs56v\" (UID: \"7a92752a-7051-4fe7-bba9-f106dd557e59\") " pod="openshift-marketplace/redhat-operators-vs56v" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.556053 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a92752a-7051-4fe7-bba9-f106dd557e59-utilities\") pod \"redhat-operators-vs56v\" (UID: \"7a92752a-7051-4fe7-bba9-f106dd557e59\") " pod="openshift-marketplace/redhat-operators-vs56v" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.556668 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a92752a-7051-4fe7-bba9-f106dd557e59-catalog-content\") pod \"redhat-operators-vs56v\" (UID: \"7a92752a-7051-4fe7-bba9-f106dd557e59\") " pod="openshift-marketplace/redhat-operators-vs56v" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.556784 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a92752a-7051-4fe7-bba9-f106dd557e59-utilities\") pod \"redhat-operators-vs56v\" (UID: \"7a92752a-7051-4fe7-bba9-f106dd557e59\") " pod="openshift-marketplace/redhat-operators-vs56v" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.577286 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5dg8\" (UniqueName: \"kubernetes.io/projected/7a92752a-7051-4fe7-bba9-f106dd557e59-kube-api-access-n5dg8\") pod \"redhat-operators-vs56v\" (UID: \"7a92752a-7051-4fe7-bba9-f106dd557e59\") " pod="openshift-marketplace/redhat-operators-vs56v" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.713128 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vs56v" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.748815 4706 scope.go:117] "RemoveContainer" containerID="8e9515fa0aa616d09ed6f6937c8190ed7aaff8a54888f168263c19f6ed4bfc5e" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.818401 4706 scope.go:117] "RemoveContainer" containerID="d97ff5efa879357f66a647a69f8a06087361f667a750a4e0db74e62cd38062b6" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.848160 4706 scope.go:117] "RemoveContainer" containerID="13808474ccfd99d7681ce4763e54a102d84d3cc2e4965084e152229020c4fea1" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.883355 4706 scope.go:117] "RemoveContainer" containerID="1fdb9aa35a8f80efeebee1cfff0f12032e090d8f7195dc70cc457a9fac504338" Dec 06 14:21:59 crc kubenswrapper[4706]: I1206 14:21:59.931957 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vs56v"] Dec 06 14:21:59 crc kubenswrapper[4706]: W1206 14:21:59.941876 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a92752a_7051_4fe7_bba9_f106dd557e59.slice/crio-1a96a8396e7ecb53a8fd043a735c769dc8f1a5b8fc3661014084911aed676967 WatchSource:0}: Error finding container 1a96a8396e7ecb53a8fd043a735c769dc8f1a5b8fc3661014084911aed676967: Status 404 returned error can't find the container with id 1a96a8396e7ecb53a8fd043a735c769dc8f1a5b8fc3661014084911aed676967 Dec 06 14:22:00 crc kubenswrapper[4706]: E1206 14:22:00.235580 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a92752a_7051_4fe7_bba9_f106dd557e59.slice/crio-conmon-fcd78d967b1df5b762c67fc583225623e3a9a192287c8cc43a9a980592a3d6c6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a92752a_7051_4fe7_bba9_f106dd557e59.slice/crio-fcd78d967b1df5b762c67fc583225623e3a9a192287c8cc43a9a980592a3d6c6.scope\": RecentStats: unable to find data in memory cache]" Dec 06 14:22:00 crc kubenswrapper[4706]: I1206 14:22:00.431162 4706 generic.go:334] "Generic (PLEG): container finished" podID="7a92752a-7051-4fe7-bba9-f106dd557e59" containerID="fcd78d967b1df5b762c67fc583225623e3a9a192287c8cc43a9a980592a3d6c6" exitCode=0 Dec 06 14:22:00 crc kubenswrapper[4706]: I1206 14:22:00.431257 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vs56v" event={"ID":"7a92752a-7051-4fe7-bba9-f106dd557e59","Type":"ContainerDied","Data":"fcd78d967b1df5b762c67fc583225623e3a9a192287c8cc43a9a980592a3d6c6"} Dec 06 14:22:00 crc kubenswrapper[4706]: I1206 14:22:00.431467 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vs56v" event={"ID":"7a92752a-7051-4fe7-bba9-f106dd557e59","Type":"ContainerStarted","Data":"1a96a8396e7ecb53a8fd043a735c769dc8f1a5b8fc3661014084911aed676967"} Dec 06 14:22:00 crc kubenswrapper[4706]: I1206 14:22:00.434674 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 14:22:01 crc kubenswrapper[4706]: I1206 14:22:01.441738 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vs56v" event={"ID":"7a92752a-7051-4fe7-bba9-f106dd557e59","Type":"ContainerStarted","Data":"c8487565d5bf0c27e508f8deb509f16e3d29ee1b146d59f626138bf0d5f12a47"} Dec 06 14:22:02 crc kubenswrapper[4706]: I1206 14:22:02.452175 4706 generic.go:334] "Generic (PLEG): container finished" podID="7a92752a-7051-4fe7-bba9-f106dd557e59" containerID="c8487565d5bf0c27e508f8deb509f16e3d29ee1b146d59f626138bf0d5f12a47" exitCode=0 Dec 06 14:22:02 crc kubenswrapper[4706]: I1206 14:22:02.452230 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vs56v" event={"ID":"7a92752a-7051-4fe7-bba9-f106dd557e59","Type":"ContainerDied","Data":"c8487565d5bf0c27e508f8deb509f16e3d29ee1b146d59f626138bf0d5f12a47"} Dec 06 14:22:03 crc kubenswrapper[4706]: I1206 14:22:03.465284 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vs56v" event={"ID":"7a92752a-7051-4fe7-bba9-f106dd557e59","Type":"ContainerStarted","Data":"225ae07651fc6b54aa55a74e79c7b1d3f318347d347f850418da0f84529cdb5c"} Dec 06 14:22:03 crc kubenswrapper[4706]: I1206 14:22:03.501359 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vs56v" podStartSLOduration=2.031004683 podStartE2EDuration="4.50133466s" podCreationTimestamp="2025-12-06 14:21:59 +0000 UTC" firstStartedPulling="2025-12-06 14:22:00.434455546 +0000 UTC m=+1563.270248403" lastFinishedPulling="2025-12-06 14:22:02.904785483 +0000 UTC m=+1565.740578380" observedRunningTime="2025-12-06 14:22:03.495609298 +0000 UTC m=+1566.331402215" watchObservedRunningTime="2025-12-06 14:22:03.50133466 +0000 UTC m=+1566.337127547" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.017057 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zwm4d/must-gather-skpfz"] Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.018424 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zwm4d/must-gather-skpfz" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.023419 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-zwm4d"/"default-dockercfg-67fgb" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.023721 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zwm4d"/"kube-root-ca.crt" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.024112 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zwm4d"/"openshift-service-ca.crt" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.033951 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zwm4d/must-gather-skpfz"] Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.090197 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b893ca08-355d-41e9-a9a1-18cceff55185-must-gather-output\") pod \"must-gather-skpfz\" (UID: \"b893ca08-355d-41e9-a9a1-18cceff55185\") " pod="openshift-must-gather-zwm4d/must-gather-skpfz" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.090263 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffjd2\" (UniqueName: \"kubernetes.io/projected/b893ca08-355d-41e9-a9a1-18cceff55185-kube-api-access-ffjd2\") pod \"must-gather-skpfz\" (UID: \"b893ca08-355d-41e9-a9a1-18cceff55185\") " pod="openshift-must-gather-zwm4d/must-gather-skpfz" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.191893 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffjd2\" (UniqueName: \"kubernetes.io/projected/b893ca08-355d-41e9-a9a1-18cceff55185-kube-api-access-ffjd2\") pod \"must-gather-skpfz\" (UID: \"b893ca08-355d-41e9-a9a1-18cceff55185\") " pod="openshift-must-gather-zwm4d/must-gather-skpfz" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.192004 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b893ca08-355d-41e9-a9a1-18cceff55185-must-gather-output\") pod \"must-gather-skpfz\" (UID: \"b893ca08-355d-41e9-a9a1-18cceff55185\") " pod="openshift-must-gather-zwm4d/must-gather-skpfz" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.192488 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b893ca08-355d-41e9-a9a1-18cceff55185-must-gather-output\") pod \"must-gather-skpfz\" (UID: \"b893ca08-355d-41e9-a9a1-18cceff55185\") " pod="openshift-must-gather-zwm4d/must-gather-skpfz" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.213223 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffjd2\" (UniqueName: \"kubernetes.io/projected/b893ca08-355d-41e9-a9a1-18cceff55185-kube-api-access-ffjd2\") pod \"must-gather-skpfz\" (UID: \"b893ca08-355d-41e9-a9a1-18cceff55185\") " pod="openshift-must-gather-zwm4d/must-gather-skpfz" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.336788 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zwm4d/must-gather-skpfz" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.547103 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zwm4d/must-gather-skpfz"] Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.713844 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vs56v" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.713904 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vs56v" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.775786 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vs56v" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.787555 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hz85m"] Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.797683 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hz85m" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.820389 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hz85m"] Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.900860 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55b482a7-1bdf-410d-825a-e976999f4de3-utilities\") pod \"redhat-marketplace-hz85m\" (UID: \"55b482a7-1bdf-410d-825a-e976999f4de3\") " pod="openshift-marketplace/redhat-marketplace-hz85m" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.900920 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55b482a7-1bdf-410d-825a-e976999f4de3-catalog-content\") pod \"redhat-marketplace-hz85m\" (UID: \"55b482a7-1bdf-410d-825a-e976999f4de3\") " pod="openshift-marketplace/redhat-marketplace-hz85m" Dec 06 14:22:09 crc kubenswrapper[4706]: I1206 14:22:09.900962 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcw4l\" (UniqueName: \"kubernetes.io/projected/55b482a7-1bdf-410d-825a-e976999f4de3-kube-api-access-qcw4l\") pod \"redhat-marketplace-hz85m\" (UID: \"55b482a7-1bdf-410d-825a-e976999f4de3\") " pod="openshift-marketplace/redhat-marketplace-hz85m" Dec 06 14:22:10 crc kubenswrapper[4706]: I1206 14:22:10.002115 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55b482a7-1bdf-410d-825a-e976999f4de3-utilities\") pod \"redhat-marketplace-hz85m\" (UID: \"55b482a7-1bdf-410d-825a-e976999f4de3\") " pod="openshift-marketplace/redhat-marketplace-hz85m" Dec 06 14:22:10 crc kubenswrapper[4706]: I1206 14:22:10.002166 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55b482a7-1bdf-410d-825a-e976999f4de3-catalog-content\") pod \"redhat-marketplace-hz85m\" (UID: \"55b482a7-1bdf-410d-825a-e976999f4de3\") " pod="openshift-marketplace/redhat-marketplace-hz85m" Dec 06 14:22:10 crc kubenswrapper[4706]: I1206 14:22:10.002190 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcw4l\" (UniqueName: \"kubernetes.io/projected/55b482a7-1bdf-410d-825a-e976999f4de3-kube-api-access-qcw4l\") pod \"redhat-marketplace-hz85m\" (UID: \"55b482a7-1bdf-410d-825a-e976999f4de3\") " pod="openshift-marketplace/redhat-marketplace-hz85m" Dec 06 14:22:10 crc kubenswrapper[4706]: I1206 14:22:10.002924 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55b482a7-1bdf-410d-825a-e976999f4de3-utilities\") pod \"redhat-marketplace-hz85m\" (UID: \"55b482a7-1bdf-410d-825a-e976999f4de3\") " pod="openshift-marketplace/redhat-marketplace-hz85m" Dec 06 14:22:10 crc kubenswrapper[4706]: I1206 14:22:10.003126 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55b482a7-1bdf-410d-825a-e976999f4de3-catalog-content\") pod \"redhat-marketplace-hz85m\" (UID: \"55b482a7-1bdf-410d-825a-e976999f4de3\") " pod="openshift-marketplace/redhat-marketplace-hz85m" Dec 06 14:22:10 crc kubenswrapper[4706]: I1206 14:22:10.023358 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcw4l\" (UniqueName: \"kubernetes.io/projected/55b482a7-1bdf-410d-825a-e976999f4de3-kube-api-access-qcw4l\") pod \"redhat-marketplace-hz85m\" (UID: \"55b482a7-1bdf-410d-825a-e976999f4de3\") " pod="openshift-marketplace/redhat-marketplace-hz85m" Dec 06 14:22:10 crc kubenswrapper[4706]: I1206 14:22:10.123484 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hz85m" Dec 06 14:22:10 crc kubenswrapper[4706]: I1206 14:22:10.367282 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hz85m"] Dec 06 14:22:10 crc kubenswrapper[4706]: I1206 14:22:10.510088 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zwm4d/must-gather-skpfz" event={"ID":"b893ca08-355d-41e9-a9a1-18cceff55185","Type":"ContainerStarted","Data":"4c9d04651e818fa57bf3bcc11e84f04c4daf2a44a8bfafbccf2c3c9f93b55795"} Dec 06 14:22:10 crc kubenswrapper[4706]: I1206 14:22:10.510130 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zwm4d/must-gather-skpfz" event={"ID":"b893ca08-355d-41e9-a9a1-18cceff55185","Type":"ContainerStarted","Data":"cb54c9e5cec94e1c193edea9567f6bd3ab1ea10c516be0d27a12e17b924eeabf"} Dec 06 14:22:10 crc kubenswrapper[4706]: I1206 14:22:10.510140 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zwm4d/must-gather-skpfz" event={"ID":"b893ca08-355d-41e9-a9a1-18cceff55185","Type":"ContainerStarted","Data":"5cde6ff85bed17fdaf632230821758e8cd8ec23885c251dfdd01061edc978eec"} Dec 06 14:22:10 crc kubenswrapper[4706]: I1206 14:22:10.511568 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hz85m" event={"ID":"55b482a7-1bdf-410d-825a-e976999f4de3","Type":"ContainerStarted","Data":"5179737468e3a3ee801df80de04b111012bce9e078ddc45bb0b2fc64533d138b"} Dec 06 14:22:10 crc kubenswrapper[4706]: I1206 14:22:10.542790 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zwm4d/must-gather-skpfz" podStartSLOduration=2.542776002 podStartE2EDuration="2.542776002s" podCreationTimestamp="2025-12-06 14:22:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 14:22:10.542235408 +0000 UTC m=+1573.378028265" watchObservedRunningTime="2025-12-06 14:22:10.542776002 +0000 UTC m=+1573.378568849" Dec 06 14:22:10 crc kubenswrapper[4706]: I1206 14:22:10.574634 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vs56v" Dec 06 14:22:11 crc kubenswrapper[4706]: I1206 14:22:11.518758 4706 generic.go:334] "Generic (PLEG): container finished" podID="55b482a7-1bdf-410d-825a-e976999f4de3" containerID="b18ce092ed7501ab7c1ed75f83f4f488db7bc86f71275457f923b6fc12404980" exitCode=0 Dec 06 14:22:11 crc kubenswrapper[4706]: I1206 14:22:11.518961 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hz85m" event={"ID":"55b482a7-1bdf-410d-825a-e976999f4de3","Type":"ContainerDied","Data":"b18ce092ed7501ab7c1ed75f83f4f488db7bc86f71275457f923b6fc12404980"} Dec 06 14:22:11 crc kubenswrapper[4706]: I1206 14:22:11.954413 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vs56v"] Dec 06 14:22:12 crc kubenswrapper[4706]: I1206 14:22:12.525648 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hz85m" event={"ID":"55b482a7-1bdf-410d-825a-e976999f4de3","Type":"ContainerStarted","Data":"735c9fb5021b8b6364d77e96639e663a55c927d050845a8c2617d05287957a4f"} Dec 06 14:22:12 crc kubenswrapper[4706]: I1206 14:22:12.525825 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vs56v" podUID="7a92752a-7051-4fe7-bba9-f106dd557e59" containerName="registry-server" containerID="cri-o://225ae07651fc6b54aa55a74e79c7b1d3f318347d347f850418da0f84529cdb5c" gracePeriod=2 Dec 06 14:22:13 crc kubenswrapper[4706]: I1206 14:22:13.538231 4706 generic.go:334] "Generic (PLEG): container finished" podID="55b482a7-1bdf-410d-825a-e976999f4de3" containerID="735c9fb5021b8b6364d77e96639e663a55c927d050845a8c2617d05287957a4f" exitCode=0 Dec 06 14:22:13 crc kubenswrapper[4706]: I1206 14:22:13.538273 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hz85m" event={"ID":"55b482a7-1bdf-410d-825a-e976999f4de3","Type":"ContainerDied","Data":"735c9fb5021b8b6364d77e96639e663a55c927d050845a8c2617d05287957a4f"} Dec 06 14:22:13 crc kubenswrapper[4706]: I1206 14:22:13.996185 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vs56v" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.055704 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a92752a-7051-4fe7-bba9-f106dd557e59-utilities\") pod \"7a92752a-7051-4fe7-bba9-f106dd557e59\" (UID: \"7a92752a-7051-4fe7-bba9-f106dd557e59\") " Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.055778 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a92752a-7051-4fe7-bba9-f106dd557e59-catalog-content\") pod \"7a92752a-7051-4fe7-bba9-f106dd557e59\" (UID: \"7a92752a-7051-4fe7-bba9-f106dd557e59\") " Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.055905 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5dg8\" (UniqueName: \"kubernetes.io/projected/7a92752a-7051-4fe7-bba9-f106dd557e59-kube-api-access-n5dg8\") pod \"7a92752a-7051-4fe7-bba9-f106dd557e59\" (UID: \"7a92752a-7051-4fe7-bba9-f106dd557e59\") " Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.056911 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a92752a-7051-4fe7-bba9-f106dd557e59-utilities" (OuterVolumeSpecName: "utilities") pod "7a92752a-7051-4fe7-bba9-f106dd557e59" (UID: "7a92752a-7051-4fe7-bba9-f106dd557e59"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.061043 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a92752a-7051-4fe7-bba9-f106dd557e59-kube-api-access-n5dg8" (OuterVolumeSpecName: "kube-api-access-n5dg8") pod "7a92752a-7051-4fe7-bba9-f106dd557e59" (UID: "7a92752a-7051-4fe7-bba9-f106dd557e59"). InnerVolumeSpecName "kube-api-access-n5dg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.157115 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5dg8\" (UniqueName: \"kubernetes.io/projected/7a92752a-7051-4fe7-bba9-f106dd557e59-kube-api-access-n5dg8\") on node \"crc\" DevicePath \"\"" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.157165 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a92752a-7051-4fe7-bba9-f106dd557e59-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.198800 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a92752a-7051-4fe7-bba9-f106dd557e59-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a92752a-7051-4fe7-bba9-f106dd557e59" (UID: "7a92752a-7051-4fe7-bba9-f106dd557e59"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.258007 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a92752a-7051-4fe7-bba9-f106dd557e59-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.548967 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hz85m" event={"ID":"55b482a7-1bdf-410d-825a-e976999f4de3","Type":"ContainerStarted","Data":"c125287324e035c499c31cc566e2856e7664180a210c1118b99364a617b7f953"} Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.551538 4706 generic.go:334] "Generic (PLEG): container finished" podID="7a92752a-7051-4fe7-bba9-f106dd557e59" containerID="225ae07651fc6b54aa55a74e79c7b1d3f318347d347f850418da0f84529cdb5c" exitCode=0 Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.551610 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vs56v" event={"ID":"7a92752a-7051-4fe7-bba9-f106dd557e59","Type":"ContainerDied","Data":"225ae07651fc6b54aa55a74e79c7b1d3f318347d347f850418da0f84529cdb5c"} Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.551689 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vs56v" event={"ID":"7a92752a-7051-4fe7-bba9-f106dd557e59","Type":"ContainerDied","Data":"1a96a8396e7ecb53a8fd043a735c769dc8f1a5b8fc3661014084911aed676967"} Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.551635 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vs56v" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.551718 4706 scope.go:117] "RemoveContainer" containerID="225ae07651fc6b54aa55a74e79c7b1d3f318347d347f850418da0f84529cdb5c" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.572341 4706 scope.go:117] "RemoveContainer" containerID="c8487565d5bf0c27e508f8deb509f16e3d29ee1b146d59f626138bf0d5f12a47" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.580062 4706 patch_prober.go:28] interesting pod/machine-config-daemon-67v56 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.580145 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.580211 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-67v56" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.581080 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f"} pod="openshift-machine-config-operator/machine-config-daemon-67v56" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.581198 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerName="machine-config-daemon" containerID="cri-o://2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" gracePeriod=600 Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.599932 4706 scope.go:117] "RemoveContainer" containerID="fcd78d967b1df5b762c67fc583225623e3a9a192287c8cc43a9a980592a3d6c6" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.605979 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hz85m" podStartSLOduration=3.147120815 podStartE2EDuration="5.605963477s" podCreationTimestamp="2025-12-06 14:22:09 +0000 UTC" firstStartedPulling="2025-12-06 14:22:11.520261026 +0000 UTC m=+1574.356053893" lastFinishedPulling="2025-12-06 14:22:13.979103698 +0000 UTC m=+1576.814896555" observedRunningTime="2025-12-06 14:22:14.603460171 +0000 UTC m=+1577.439253068" watchObservedRunningTime="2025-12-06 14:22:14.605963477 +0000 UTC m=+1577.441756334" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.622351 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vs56v"] Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.626524 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vs56v"] Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.638762 4706 scope.go:117] "RemoveContainer" containerID="225ae07651fc6b54aa55a74e79c7b1d3f318347d347f850418da0f84529cdb5c" Dec 06 14:22:14 crc kubenswrapper[4706]: E1206 14:22:14.641657 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"225ae07651fc6b54aa55a74e79c7b1d3f318347d347f850418da0f84529cdb5c\": container with ID starting with 225ae07651fc6b54aa55a74e79c7b1d3f318347d347f850418da0f84529cdb5c not found: ID does not exist" containerID="225ae07651fc6b54aa55a74e79c7b1d3f318347d347f850418da0f84529cdb5c" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.641698 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"225ae07651fc6b54aa55a74e79c7b1d3f318347d347f850418da0f84529cdb5c"} err="failed to get container status \"225ae07651fc6b54aa55a74e79c7b1d3f318347d347f850418da0f84529cdb5c\": rpc error: code = NotFound desc = could not find container \"225ae07651fc6b54aa55a74e79c7b1d3f318347d347f850418da0f84529cdb5c\": container with ID starting with 225ae07651fc6b54aa55a74e79c7b1d3f318347d347f850418da0f84529cdb5c not found: ID does not exist" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.641729 4706 scope.go:117] "RemoveContainer" containerID="c8487565d5bf0c27e508f8deb509f16e3d29ee1b146d59f626138bf0d5f12a47" Dec 06 14:22:14 crc kubenswrapper[4706]: E1206 14:22:14.642256 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8487565d5bf0c27e508f8deb509f16e3d29ee1b146d59f626138bf0d5f12a47\": container with ID starting with c8487565d5bf0c27e508f8deb509f16e3d29ee1b146d59f626138bf0d5f12a47 not found: ID does not exist" containerID="c8487565d5bf0c27e508f8deb509f16e3d29ee1b146d59f626138bf0d5f12a47" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.642360 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8487565d5bf0c27e508f8deb509f16e3d29ee1b146d59f626138bf0d5f12a47"} err="failed to get container status \"c8487565d5bf0c27e508f8deb509f16e3d29ee1b146d59f626138bf0d5f12a47\": rpc error: code = NotFound desc = could not find container \"c8487565d5bf0c27e508f8deb509f16e3d29ee1b146d59f626138bf0d5f12a47\": container with ID starting with c8487565d5bf0c27e508f8deb509f16e3d29ee1b146d59f626138bf0d5f12a47 not found: ID does not exist" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.642414 4706 scope.go:117] "RemoveContainer" containerID="fcd78d967b1df5b762c67fc583225623e3a9a192287c8cc43a9a980592a3d6c6" Dec 06 14:22:14 crc kubenswrapper[4706]: E1206 14:22:14.642759 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcd78d967b1df5b762c67fc583225623e3a9a192287c8cc43a9a980592a3d6c6\": container with ID starting with fcd78d967b1df5b762c67fc583225623e3a9a192287c8cc43a9a980592a3d6c6 not found: ID does not exist" containerID="fcd78d967b1df5b762c67fc583225623e3a9a192287c8cc43a9a980592a3d6c6" Dec 06 14:22:14 crc kubenswrapper[4706]: I1206 14:22:14.642790 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcd78d967b1df5b762c67fc583225623e3a9a192287c8cc43a9a980592a3d6c6"} err="failed to get container status \"fcd78d967b1df5b762c67fc583225623e3a9a192287c8cc43a9a980592a3d6c6\": rpc error: code = NotFound desc = could not find container \"fcd78d967b1df5b762c67fc583225623e3a9a192287c8cc43a9a980592a3d6c6\": container with ID starting with fcd78d967b1df5b762c67fc583225623e3a9a192287c8cc43a9a980592a3d6c6 not found: ID does not exist" Dec 06 14:22:14 crc kubenswrapper[4706]: E1206 14:22:14.720734 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:22:15 crc kubenswrapper[4706]: I1206 14:22:15.518687 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a92752a-7051-4fe7-bba9-f106dd557e59" path="/var/lib/kubelet/pods/7a92752a-7051-4fe7-bba9-f106dd557e59/volumes" Dec 06 14:22:15 crc kubenswrapper[4706]: I1206 14:22:15.557146 4706 generic.go:334] "Generic (PLEG): container finished" podID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" exitCode=0 Dec 06 14:22:15 crc kubenswrapper[4706]: I1206 14:22:15.557198 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerDied","Data":"2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f"} Dec 06 14:22:15 crc kubenswrapper[4706]: I1206 14:22:15.557228 4706 scope.go:117] "RemoveContainer" containerID="541fe7945667e62148e85f3104c224b6e7d1a0ff6a7b9516e33ef2027678744f" Dec 06 14:22:15 crc kubenswrapper[4706]: I1206 14:22:15.557666 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:22:15 crc kubenswrapper[4706]: E1206 14:22:15.557858 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:22:20 crc kubenswrapper[4706]: I1206 14:22:20.124393 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hz85m" Dec 06 14:22:20 crc kubenswrapper[4706]: I1206 14:22:20.124753 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hz85m" Dec 06 14:22:20 crc kubenswrapper[4706]: I1206 14:22:20.171363 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hz85m" Dec 06 14:22:20 crc kubenswrapper[4706]: I1206 14:22:20.671225 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hz85m" Dec 06 14:22:20 crc kubenswrapper[4706]: I1206 14:22:20.744656 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hz85m"] Dec 06 14:22:22 crc kubenswrapper[4706]: I1206 14:22:22.610836 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hz85m" podUID="55b482a7-1bdf-410d-825a-e976999f4de3" containerName="registry-server" containerID="cri-o://c125287324e035c499c31cc566e2856e7664180a210c1118b99364a617b7f953" gracePeriod=2 Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.473065 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hz85m" Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.577870 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcw4l\" (UniqueName: \"kubernetes.io/projected/55b482a7-1bdf-410d-825a-e976999f4de3-kube-api-access-qcw4l\") pod \"55b482a7-1bdf-410d-825a-e976999f4de3\" (UID: \"55b482a7-1bdf-410d-825a-e976999f4de3\") " Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.578020 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55b482a7-1bdf-410d-825a-e976999f4de3-catalog-content\") pod \"55b482a7-1bdf-410d-825a-e976999f4de3\" (UID: \"55b482a7-1bdf-410d-825a-e976999f4de3\") " Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.578055 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55b482a7-1bdf-410d-825a-e976999f4de3-utilities\") pod \"55b482a7-1bdf-410d-825a-e976999f4de3\" (UID: \"55b482a7-1bdf-410d-825a-e976999f4de3\") " Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.579382 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55b482a7-1bdf-410d-825a-e976999f4de3-utilities" (OuterVolumeSpecName: "utilities") pod "55b482a7-1bdf-410d-825a-e976999f4de3" (UID: "55b482a7-1bdf-410d-825a-e976999f4de3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.589464 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55b482a7-1bdf-410d-825a-e976999f4de3-kube-api-access-qcw4l" (OuterVolumeSpecName: "kube-api-access-qcw4l") pod "55b482a7-1bdf-410d-825a-e976999f4de3" (UID: "55b482a7-1bdf-410d-825a-e976999f4de3"). InnerVolumeSpecName "kube-api-access-qcw4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.600136 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55b482a7-1bdf-410d-825a-e976999f4de3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55b482a7-1bdf-410d-825a-e976999f4de3" (UID: "55b482a7-1bdf-410d-825a-e976999f4de3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.617933 4706 generic.go:334] "Generic (PLEG): container finished" podID="55b482a7-1bdf-410d-825a-e976999f4de3" containerID="c125287324e035c499c31cc566e2856e7664180a210c1118b99364a617b7f953" exitCode=0 Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.617979 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hz85m" event={"ID":"55b482a7-1bdf-410d-825a-e976999f4de3","Type":"ContainerDied","Data":"c125287324e035c499c31cc566e2856e7664180a210c1118b99364a617b7f953"} Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.617994 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hz85m" Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.618011 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hz85m" event={"ID":"55b482a7-1bdf-410d-825a-e976999f4de3","Type":"ContainerDied","Data":"5179737468e3a3ee801df80de04b111012bce9e078ddc45bb0b2fc64533d138b"} Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.618031 4706 scope.go:117] "RemoveContainer" containerID="c125287324e035c499c31cc566e2856e7664180a210c1118b99364a617b7f953" Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.648424 4706 scope.go:117] "RemoveContainer" containerID="735c9fb5021b8b6364d77e96639e663a55c927d050845a8c2617d05287957a4f" Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.649974 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hz85m"] Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.656491 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hz85m"] Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.663997 4706 scope.go:117] "RemoveContainer" containerID="b18ce092ed7501ab7c1ed75f83f4f488db7bc86f71275457f923b6fc12404980" Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.679440 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcw4l\" (UniqueName: \"kubernetes.io/projected/55b482a7-1bdf-410d-825a-e976999f4de3-kube-api-access-qcw4l\") on node \"crc\" DevicePath \"\"" Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.679472 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55b482a7-1bdf-410d-825a-e976999f4de3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.679484 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55b482a7-1bdf-410d-825a-e976999f4de3-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.696630 4706 scope.go:117] "RemoveContainer" containerID="c125287324e035c499c31cc566e2856e7664180a210c1118b99364a617b7f953" Dec 06 14:22:23 crc kubenswrapper[4706]: E1206 14:22:23.697235 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c125287324e035c499c31cc566e2856e7664180a210c1118b99364a617b7f953\": container with ID starting with c125287324e035c499c31cc566e2856e7664180a210c1118b99364a617b7f953 not found: ID does not exist" containerID="c125287324e035c499c31cc566e2856e7664180a210c1118b99364a617b7f953" Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.697283 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c125287324e035c499c31cc566e2856e7664180a210c1118b99364a617b7f953"} err="failed to get container status \"c125287324e035c499c31cc566e2856e7664180a210c1118b99364a617b7f953\": rpc error: code = NotFound desc = could not find container \"c125287324e035c499c31cc566e2856e7664180a210c1118b99364a617b7f953\": container with ID starting with c125287324e035c499c31cc566e2856e7664180a210c1118b99364a617b7f953 not found: ID does not exist" Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.697359 4706 scope.go:117] "RemoveContainer" containerID="735c9fb5021b8b6364d77e96639e663a55c927d050845a8c2617d05287957a4f" Dec 06 14:22:23 crc kubenswrapper[4706]: E1206 14:22:23.697643 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"735c9fb5021b8b6364d77e96639e663a55c927d050845a8c2617d05287957a4f\": container with ID starting with 735c9fb5021b8b6364d77e96639e663a55c927d050845a8c2617d05287957a4f not found: ID does not exist" containerID="735c9fb5021b8b6364d77e96639e663a55c927d050845a8c2617d05287957a4f" Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.697669 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"735c9fb5021b8b6364d77e96639e663a55c927d050845a8c2617d05287957a4f"} err="failed to get container status \"735c9fb5021b8b6364d77e96639e663a55c927d050845a8c2617d05287957a4f\": rpc error: code = NotFound desc = could not find container \"735c9fb5021b8b6364d77e96639e663a55c927d050845a8c2617d05287957a4f\": container with ID starting with 735c9fb5021b8b6364d77e96639e663a55c927d050845a8c2617d05287957a4f not found: ID does not exist" Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.697685 4706 scope.go:117] "RemoveContainer" containerID="b18ce092ed7501ab7c1ed75f83f4f488db7bc86f71275457f923b6fc12404980" Dec 06 14:22:23 crc kubenswrapper[4706]: E1206 14:22:23.697900 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b18ce092ed7501ab7c1ed75f83f4f488db7bc86f71275457f923b6fc12404980\": container with ID starting with b18ce092ed7501ab7c1ed75f83f4f488db7bc86f71275457f923b6fc12404980 not found: ID does not exist" containerID="b18ce092ed7501ab7c1ed75f83f4f488db7bc86f71275457f923b6fc12404980" Dec 06 14:22:23 crc kubenswrapper[4706]: I1206 14:22:23.697925 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b18ce092ed7501ab7c1ed75f83f4f488db7bc86f71275457f923b6fc12404980"} err="failed to get container status \"b18ce092ed7501ab7c1ed75f83f4f488db7bc86f71275457f923b6fc12404980\": rpc error: code = NotFound desc = could not find container \"b18ce092ed7501ab7c1ed75f83f4f488db7bc86f71275457f923b6fc12404980\": container with ID starting with b18ce092ed7501ab7c1ed75f83f4f488db7bc86f71275457f923b6fc12404980 not found: ID does not exist" Dec 06 14:22:24 crc kubenswrapper[4706]: I1206 14:22:24.955620 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xgl8w"] Dec 06 14:22:24 crc kubenswrapper[4706]: E1206 14:22:24.955824 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55b482a7-1bdf-410d-825a-e976999f4de3" containerName="extract-utilities" Dec 06 14:22:24 crc kubenswrapper[4706]: I1206 14:22:24.955835 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="55b482a7-1bdf-410d-825a-e976999f4de3" containerName="extract-utilities" Dec 06 14:22:24 crc kubenswrapper[4706]: E1206 14:22:24.955844 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55b482a7-1bdf-410d-825a-e976999f4de3" containerName="extract-content" Dec 06 14:22:24 crc kubenswrapper[4706]: I1206 14:22:24.955851 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="55b482a7-1bdf-410d-825a-e976999f4de3" containerName="extract-content" Dec 06 14:22:24 crc kubenswrapper[4706]: E1206 14:22:24.955864 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a92752a-7051-4fe7-bba9-f106dd557e59" containerName="extract-content" Dec 06 14:22:24 crc kubenswrapper[4706]: I1206 14:22:24.955871 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a92752a-7051-4fe7-bba9-f106dd557e59" containerName="extract-content" Dec 06 14:22:24 crc kubenswrapper[4706]: E1206 14:22:24.955878 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55b482a7-1bdf-410d-825a-e976999f4de3" containerName="registry-server" Dec 06 14:22:24 crc kubenswrapper[4706]: I1206 14:22:24.955883 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="55b482a7-1bdf-410d-825a-e976999f4de3" containerName="registry-server" Dec 06 14:22:24 crc kubenswrapper[4706]: E1206 14:22:24.955896 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a92752a-7051-4fe7-bba9-f106dd557e59" containerName="extract-utilities" Dec 06 14:22:24 crc kubenswrapper[4706]: I1206 14:22:24.955902 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a92752a-7051-4fe7-bba9-f106dd557e59" containerName="extract-utilities" Dec 06 14:22:24 crc kubenswrapper[4706]: E1206 14:22:24.955910 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a92752a-7051-4fe7-bba9-f106dd557e59" containerName="registry-server" Dec 06 14:22:24 crc kubenswrapper[4706]: I1206 14:22:24.955916 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a92752a-7051-4fe7-bba9-f106dd557e59" containerName="registry-server" Dec 06 14:22:24 crc kubenswrapper[4706]: I1206 14:22:24.956024 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="55b482a7-1bdf-410d-825a-e976999f4de3" containerName="registry-server" Dec 06 14:22:24 crc kubenswrapper[4706]: I1206 14:22:24.956037 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a92752a-7051-4fe7-bba9-f106dd557e59" containerName="registry-server" Dec 06 14:22:24 crc kubenswrapper[4706]: I1206 14:22:24.956955 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xgl8w" Dec 06 14:22:24 crc kubenswrapper[4706]: I1206 14:22:24.974999 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xgl8w"] Dec 06 14:22:24 crc kubenswrapper[4706]: I1206 14:22:24.997338 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca8afca-8dd4-475a-8f47-d492f2940eee-utilities\") pod \"certified-operators-xgl8w\" (UID: \"0ca8afca-8dd4-475a-8f47-d492f2940eee\") " pod="openshift-marketplace/certified-operators-xgl8w" Dec 06 14:22:24 crc kubenswrapper[4706]: I1206 14:22:24.997631 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r66x4\" (UniqueName: \"kubernetes.io/projected/0ca8afca-8dd4-475a-8f47-d492f2940eee-kube-api-access-r66x4\") pod \"certified-operators-xgl8w\" (UID: \"0ca8afca-8dd4-475a-8f47-d492f2940eee\") " pod="openshift-marketplace/certified-operators-xgl8w" Dec 06 14:22:24 crc kubenswrapper[4706]: I1206 14:22:24.997728 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca8afca-8dd4-475a-8f47-d492f2940eee-catalog-content\") pod \"certified-operators-xgl8w\" (UID: \"0ca8afca-8dd4-475a-8f47-d492f2940eee\") " pod="openshift-marketplace/certified-operators-xgl8w" Dec 06 14:22:25 crc kubenswrapper[4706]: I1206 14:22:25.099218 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca8afca-8dd4-475a-8f47-d492f2940eee-utilities\") pod \"certified-operators-xgl8w\" (UID: \"0ca8afca-8dd4-475a-8f47-d492f2940eee\") " pod="openshift-marketplace/certified-operators-xgl8w" Dec 06 14:22:25 crc kubenswrapper[4706]: I1206 14:22:25.099293 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r66x4\" (UniqueName: \"kubernetes.io/projected/0ca8afca-8dd4-475a-8f47-d492f2940eee-kube-api-access-r66x4\") pod \"certified-operators-xgl8w\" (UID: \"0ca8afca-8dd4-475a-8f47-d492f2940eee\") " pod="openshift-marketplace/certified-operators-xgl8w" Dec 06 14:22:25 crc kubenswrapper[4706]: I1206 14:22:25.099333 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca8afca-8dd4-475a-8f47-d492f2940eee-catalog-content\") pod \"certified-operators-xgl8w\" (UID: \"0ca8afca-8dd4-475a-8f47-d492f2940eee\") " pod="openshift-marketplace/certified-operators-xgl8w" Dec 06 14:22:25 crc kubenswrapper[4706]: I1206 14:22:25.099780 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca8afca-8dd4-475a-8f47-d492f2940eee-catalog-content\") pod \"certified-operators-xgl8w\" (UID: \"0ca8afca-8dd4-475a-8f47-d492f2940eee\") " pod="openshift-marketplace/certified-operators-xgl8w" Dec 06 14:22:25 crc kubenswrapper[4706]: I1206 14:22:25.099779 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca8afca-8dd4-475a-8f47-d492f2940eee-utilities\") pod \"certified-operators-xgl8w\" (UID: \"0ca8afca-8dd4-475a-8f47-d492f2940eee\") " pod="openshift-marketplace/certified-operators-xgl8w" Dec 06 14:22:25 crc kubenswrapper[4706]: I1206 14:22:25.141348 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r66x4\" (UniqueName: \"kubernetes.io/projected/0ca8afca-8dd4-475a-8f47-d492f2940eee-kube-api-access-r66x4\") pod \"certified-operators-xgl8w\" (UID: \"0ca8afca-8dd4-475a-8f47-d492f2940eee\") " pod="openshift-marketplace/certified-operators-xgl8w" Dec 06 14:22:25 crc kubenswrapper[4706]: I1206 14:22:25.269409 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xgl8w" Dec 06 14:22:25 crc kubenswrapper[4706]: I1206 14:22:25.517992 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55b482a7-1bdf-410d-825a-e976999f4de3" path="/var/lib/kubelet/pods/55b482a7-1bdf-410d-825a-e976999f4de3/volumes" Dec 06 14:22:25 crc kubenswrapper[4706]: I1206 14:22:25.715503 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xgl8w"] Dec 06 14:22:26 crc kubenswrapper[4706]: I1206 14:22:26.649512 4706 generic.go:334] "Generic (PLEG): container finished" podID="0ca8afca-8dd4-475a-8f47-d492f2940eee" containerID="513f5a03fa950ddceef051c442816bd127fe77866c7ba381608a8cf7ad098d57" exitCode=0 Dec 06 14:22:26 crc kubenswrapper[4706]: I1206 14:22:26.649567 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgl8w" event={"ID":"0ca8afca-8dd4-475a-8f47-d492f2940eee","Type":"ContainerDied","Data":"513f5a03fa950ddceef051c442816bd127fe77866c7ba381608a8cf7ad098d57"} Dec 06 14:22:26 crc kubenswrapper[4706]: I1206 14:22:26.649821 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgl8w" event={"ID":"0ca8afca-8dd4-475a-8f47-d492f2940eee","Type":"ContainerStarted","Data":"d30bcee84869b70e55280a2c8fa18ad4055d18301962366d92df489870581442"} Dec 06 14:22:28 crc kubenswrapper[4706]: I1206 14:22:28.664843 4706 generic.go:334] "Generic (PLEG): container finished" podID="0ca8afca-8dd4-475a-8f47-d492f2940eee" containerID="d851afcdcec468257dba99796f3646ae00314cd5d13ada20ab887bb3e1e85e7d" exitCode=0 Dec 06 14:22:28 crc kubenswrapper[4706]: I1206 14:22:28.664890 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgl8w" event={"ID":"0ca8afca-8dd4-475a-8f47-d492f2940eee","Type":"ContainerDied","Data":"d851afcdcec468257dba99796f3646ae00314cd5d13ada20ab887bb3e1e85e7d"} Dec 06 14:22:29 crc kubenswrapper[4706]: I1206 14:22:29.672201 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgl8w" event={"ID":"0ca8afca-8dd4-475a-8f47-d492f2940eee","Type":"ContainerStarted","Data":"7424be90a17c18fbbc583b980cd6b61fb4ceda13c5776c68a16775590d14eb1c"} Dec 06 14:22:29 crc kubenswrapper[4706]: I1206 14:22:29.691911 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xgl8w" podStartSLOduration=3.099782892 podStartE2EDuration="5.691890443s" podCreationTimestamp="2025-12-06 14:22:24 +0000 UTC" firstStartedPulling="2025-12-06 14:22:26.65522995 +0000 UTC m=+1589.491022797" lastFinishedPulling="2025-12-06 14:22:29.247337491 +0000 UTC m=+1592.083130348" observedRunningTime="2025-12-06 14:22:29.687133747 +0000 UTC m=+1592.522926604" watchObservedRunningTime="2025-12-06 14:22:29.691890443 +0000 UTC m=+1592.527683300" Dec 06 14:22:30 crc kubenswrapper[4706]: I1206 14:22:30.512109 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:22:30 crc kubenswrapper[4706]: E1206 14:22:30.512486 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:22:35 crc kubenswrapper[4706]: I1206 14:22:35.270452 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xgl8w" Dec 06 14:22:35 crc kubenswrapper[4706]: I1206 14:22:35.270851 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xgl8w" Dec 06 14:22:35 crc kubenswrapper[4706]: I1206 14:22:35.331822 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xgl8w" Dec 06 14:22:35 crc kubenswrapper[4706]: I1206 14:22:35.755728 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xgl8w" Dec 06 14:22:35 crc kubenswrapper[4706]: I1206 14:22:35.802710 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xgl8w"] Dec 06 14:22:37 crc kubenswrapper[4706]: I1206 14:22:37.718847 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xgl8w" podUID="0ca8afca-8dd4-475a-8f47-d492f2940eee" containerName="registry-server" containerID="cri-o://7424be90a17c18fbbc583b980cd6b61fb4ceda13c5776c68a16775590d14eb1c" gracePeriod=2 Dec 06 14:22:38 crc kubenswrapper[4706]: I1206 14:22:38.728405 4706 generic.go:334] "Generic (PLEG): container finished" podID="0ca8afca-8dd4-475a-8f47-d492f2940eee" containerID="7424be90a17c18fbbc583b980cd6b61fb4ceda13c5776c68a16775590d14eb1c" exitCode=0 Dec 06 14:22:38 crc kubenswrapper[4706]: I1206 14:22:38.728557 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgl8w" event={"ID":"0ca8afca-8dd4-475a-8f47-d492f2940eee","Type":"ContainerDied","Data":"7424be90a17c18fbbc583b980cd6b61fb4ceda13c5776c68a16775590d14eb1c"} Dec 06 14:22:38 crc kubenswrapper[4706]: I1206 14:22:38.833818 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xgl8w" Dec 06 14:22:38 crc kubenswrapper[4706]: I1206 14:22:38.995769 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca8afca-8dd4-475a-8f47-d492f2940eee-utilities\") pod \"0ca8afca-8dd4-475a-8f47-d492f2940eee\" (UID: \"0ca8afca-8dd4-475a-8f47-d492f2940eee\") " Dec 06 14:22:38 crc kubenswrapper[4706]: I1206 14:22:38.995842 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r66x4\" (UniqueName: \"kubernetes.io/projected/0ca8afca-8dd4-475a-8f47-d492f2940eee-kube-api-access-r66x4\") pod \"0ca8afca-8dd4-475a-8f47-d492f2940eee\" (UID: \"0ca8afca-8dd4-475a-8f47-d492f2940eee\") " Dec 06 14:22:38 crc kubenswrapper[4706]: I1206 14:22:38.995865 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca8afca-8dd4-475a-8f47-d492f2940eee-catalog-content\") pod \"0ca8afca-8dd4-475a-8f47-d492f2940eee\" (UID: \"0ca8afca-8dd4-475a-8f47-d492f2940eee\") " Dec 06 14:22:38 crc kubenswrapper[4706]: I1206 14:22:38.997245 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ca8afca-8dd4-475a-8f47-d492f2940eee-utilities" (OuterVolumeSpecName: "utilities") pod "0ca8afca-8dd4-475a-8f47-d492f2940eee" (UID: "0ca8afca-8dd4-475a-8f47-d492f2940eee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:22:39 crc kubenswrapper[4706]: I1206 14:22:39.005623 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ca8afca-8dd4-475a-8f47-d492f2940eee-kube-api-access-r66x4" (OuterVolumeSpecName: "kube-api-access-r66x4") pod "0ca8afca-8dd4-475a-8f47-d492f2940eee" (UID: "0ca8afca-8dd4-475a-8f47-d492f2940eee"). InnerVolumeSpecName "kube-api-access-r66x4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:22:39 crc kubenswrapper[4706]: I1206 14:22:39.046073 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ca8afca-8dd4-475a-8f47-d492f2940eee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ca8afca-8dd4-475a-8f47-d492f2940eee" (UID: "0ca8afca-8dd4-475a-8f47-d492f2940eee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:22:39 crc kubenswrapper[4706]: I1206 14:22:39.097391 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r66x4\" (UniqueName: \"kubernetes.io/projected/0ca8afca-8dd4-475a-8f47-d492f2940eee-kube-api-access-r66x4\") on node \"crc\" DevicePath \"\"" Dec 06 14:22:39 crc kubenswrapper[4706]: I1206 14:22:39.097433 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca8afca-8dd4-475a-8f47-d492f2940eee-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 14:22:39 crc kubenswrapper[4706]: I1206 14:22:39.097446 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca8afca-8dd4-475a-8f47-d492f2940eee-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 14:22:39 crc kubenswrapper[4706]: I1206 14:22:39.737521 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgl8w" event={"ID":"0ca8afca-8dd4-475a-8f47-d492f2940eee","Type":"ContainerDied","Data":"d30bcee84869b70e55280a2c8fa18ad4055d18301962366d92df489870581442"} Dec 06 14:22:39 crc kubenswrapper[4706]: I1206 14:22:39.737635 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xgl8w" Dec 06 14:22:39 crc kubenswrapper[4706]: I1206 14:22:39.737787 4706 scope.go:117] "RemoveContainer" containerID="7424be90a17c18fbbc583b980cd6b61fb4ceda13c5776c68a16775590d14eb1c" Dec 06 14:22:39 crc kubenswrapper[4706]: I1206 14:22:39.753461 4706 scope.go:117] "RemoveContainer" containerID="d851afcdcec468257dba99796f3646ae00314cd5d13ada20ab887bb3e1e85e7d" Dec 06 14:22:39 crc kubenswrapper[4706]: I1206 14:22:39.770849 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xgl8w"] Dec 06 14:22:39 crc kubenswrapper[4706]: I1206 14:22:39.772793 4706 scope.go:117] "RemoveContainer" containerID="513f5a03fa950ddceef051c442816bd127fe77866c7ba381608a8cf7ad098d57" Dec 06 14:22:39 crc kubenswrapper[4706]: I1206 14:22:39.778614 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xgl8w"] Dec 06 14:22:41 crc kubenswrapper[4706]: I1206 14:22:41.511469 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:22:41 crc kubenswrapper[4706]: E1206 14:22:41.511791 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:22:41 crc kubenswrapper[4706]: I1206 14:22:41.524591 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ca8afca-8dd4-475a-8f47-d492f2940eee" path="/var/lib/kubelet/pods/0ca8afca-8dd4-475a-8f47-d492f2940eee/volumes" Dec 06 14:22:56 crc kubenswrapper[4706]: I1206 14:22:56.511857 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:22:56 crc kubenswrapper[4706]: E1206 14:22:56.513355 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:22:59 crc kubenswrapper[4706]: I1206 14:22:59.959103 4706 scope.go:117] "RemoveContainer" containerID="dda4728304f8c40611067b964e75b88e7469efe8dde5ec2124e7df8c3daf3639" Dec 06 14:23:00 crc kubenswrapper[4706]: I1206 14:23:00.511557 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-p85f8_77b4e41c-830d-48c2-8d00-97b33c7e8a29/control-plane-machine-set-operator/0.log" Dec 06 14:23:00 crc kubenswrapper[4706]: I1206 14:23:00.641933 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9hxp8_4c02a475-2aec-4c9c-ba4d-f20059c09682/kube-rbac-proxy/0.log" Dec 06 14:23:00 crc kubenswrapper[4706]: I1206 14:23:00.656419 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9hxp8_4c02a475-2aec-4c9c-ba4d-f20059c09682/machine-api-operator/0.log" Dec 06 14:23:08 crc kubenswrapper[4706]: I1206 14:23:08.511238 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:23:08 crc kubenswrapper[4706]: E1206 14:23:08.512130 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:23:16 crc kubenswrapper[4706]: I1206 14:23:16.812354 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-bcgwr_d6c3a1b0-6512-40a8-aac5-bc6629793ab8/kube-rbac-proxy/0.log" Dec 06 14:23:16 crc kubenswrapper[4706]: I1206 14:23:16.865805 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-bcgwr_d6c3a1b0-6512-40a8-aac5-bc6629793ab8/controller/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.028527 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-5vmhj_7cd749c9-c857-4019-9fb3-f2c53f0b9be8/frr-k8s-webhook-server/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.037588 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-frr-files/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.231676 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-frr-files/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.242635 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-metrics/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.279284 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-reloader/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.302619 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-reloader/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.398745 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-frr-files/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.408966 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-reloader/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.434221 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-metrics/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.486807 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-metrics/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.608529 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-frr-files/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.613363 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-reloader/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.649747 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/cp-metrics/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.660387 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/controller/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.780946 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/frr-metrics/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.813702 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/kube-rbac-proxy/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.824536 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/kube-rbac-proxy-frr/0.log" Dec 06 14:23:17 crc kubenswrapper[4706]: I1206 14:23:17.949025 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/reloader/0.log" Dec 06 14:23:18 crc kubenswrapper[4706]: I1206 14:23:18.018211 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-548bff8fd9-mt2r9_5acf3c41-82f9-4b93-8742-1b4975610426/manager/0.log" Dec 06 14:23:18 crc kubenswrapper[4706]: I1206 14:23:18.168703 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-56bc6d5c44-tg7hm_d16f9b04-736c-4331-8ad8-58d14bd858e6/webhook-server/0.log" Dec 06 14:23:18 crc kubenswrapper[4706]: I1206 14:23:18.233099 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zffsv_cb8721ae-d9da-4caf-87d3-f2dbf09c037b/frr/0.log" Dec 06 14:23:18 crc kubenswrapper[4706]: I1206 14:23:18.343936 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-w6jgb_afd1dfce-966c-4506-bf87-3e2730857136/kube-rbac-proxy/0.log" Dec 06 14:23:18 crc kubenswrapper[4706]: I1206 14:23:18.422329 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-w6jgb_afd1dfce-966c-4506-bf87-3e2730857136/speaker/0.log" Dec 06 14:23:19 crc kubenswrapper[4706]: I1206 14:23:19.511446 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:23:19 crc kubenswrapper[4706]: E1206 14:23:19.511841 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:23:34 crc kubenswrapper[4706]: I1206 14:23:34.511420 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:23:34 crc kubenswrapper[4706]: E1206 14:23:34.512205 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:23:43 crc kubenswrapper[4706]: I1206 14:23:43.546633 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4/util/0.log" Dec 06 14:23:43 crc kubenswrapper[4706]: I1206 14:23:43.725844 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4/pull/0.log" Dec 06 14:23:43 crc kubenswrapper[4706]: I1206 14:23:43.738256 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4/util/0.log" Dec 06 14:23:43 crc kubenswrapper[4706]: I1206 14:23:43.772892 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4/pull/0.log" Dec 06 14:23:43 crc kubenswrapper[4706]: I1206 14:23:43.894097 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4/pull/0.log" Dec 06 14:23:43 crc kubenswrapper[4706]: I1206 14:23:43.905280 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4/extract/0.log" Dec 06 14:23:43 crc kubenswrapper[4706]: I1206 14:23:43.909976 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqd4s_51fa29f5-f7d7-4ac2-a15a-2b16025a00b4/util/0.log" Dec 06 14:23:44 crc kubenswrapper[4706]: I1206 14:23:44.031573 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ssx77_6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110/extract-utilities/0.log" Dec 06 14:23:44 crc kubenswrapper[4706]: I1206 14:23:44.195080 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ssx77_6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110/extract-content/0.log" Dec 06 14:23:44 crc kubenswrapper[4706]: I1206 14:23:44.200708 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ssx77_6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110/extract-content/0.log" Dec 06 14:23:44 crc kubenswrapper[4706]: I1206 14:23:44.237721 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ssx77_6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110/extract-utilities/0.log" Dec 06 14:23:44 crc kubenswrapper[4706]: I1206 14:23:44.355181 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ssx77_6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110/extract-utilities/0.log" Dec 06 14:23:44 crc kubenswrapper[4706]: I1206 14:23:44.384397 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ssx77_6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110/extract-content/0.log" Dec 06 14:23:44 crc kubenswrapper[4706]: I1206 14:23:44.577501 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ggc6c_60caaf13-e871-4029-8614-dc43c10ac358/extract-utilities/0.log" Dec 06 14:23:44 crc kubenswrapper[4706]: I1206 14:23:44.705330 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ssx77_6ea6c2f3-7a03-4e0a-9be9-be2a62a5f110/registry-server/0.log" Dec 06 14:23:44 crc kubenswrapper[4706]: I1206 14:23:44.713793 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ggc6c_60caaf13-e871-4029-8614-dc43c10ac358/extract-utilities/0.log" Dec 06 14:23:44 crc kubenswrapper[4706]: I1206 14:23:44.731906 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ggc6c_60caaf13-e871-4029-8614-dc43c10ac358/extract-content/0.log" Dec 06 14:23:44 crc kubenswrapper[4706]: I1206 14:23:44.768690 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ggc6c_60caaf13-e871-4029-8614-dc43c10ac358/extract-content/0.log" Dec 06 14:23:44 crc kubenswrapper[4706]: I1206 14:23:44.867618 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ggc6c_60caaf13-e871-4029-8614-dc43c10ac358/extract-utilities/0.log" Dec 06 14:23:44 crc kubenswrapper[4706]: I1206 14:23:44.884251 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ggc6c_60caaf13-e871-4029-8614-dc43c10ac358/extract-content/0.log" Dec 06 14:23:45 crc kubenswrapper[4706]: I1206 14:23:45.146207 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-sgfd9_92920365-098a-442a-87de-91aa1479ff1d/marketplace-operator/0.log" Dec 06 14:23:45 crc kubenswrapper[4706]: I1206 14:23:45.163038 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-62dkd_0514bb3a-f280-4208-9dff-ca5549dff0ca/extract-utilities/0.log" Dec 06 14:23:45 crc kubenswrapper[4706]: I1206 14:23:45.196844 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ggc6c_60caaf13-e871-4029-8614-dc43c10ac358/registry-server/0.log" Dec 06 14:23:45 crc kubenswrapper[4706]: I1206 14:23:45.308400 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-62dkd_0514bb3a-f280-4208-9dff-ca5549dff0ca/extract-content/0.log" Dec 06 14:23:45 crc kubenswrapper[4706]: I1206 14:23:45.325167 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-62dkd_0514bb3a-f280-4208-9dff-ca5549dff0ca/extract-content/0.log" Dec 06 14:23:45 crc kubenswrapper[4706]: I1206 14:23:45.332003 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-62dkd_0514bb3a-f280-4208-9dff-ca5549dff0ca/extract-utilities/0.log" Dec 06 14:23:45 crc kubenswrapper[4706]: I1206 14:23:45.453689 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-62dkd_0514bb3a-f280-4208-9dff-ca5549dff0ca/extract-utilities/0.log" Dec 06 14:23:45 crc kubenswrapper[4706]: I1206 14:23:45.486563 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-62dkd_0514bb3a-f280-4208-9dff-ca5549dff0ca/extract-content/0.log" Dec 06 14:23:45 crc kubenswrapper[4706]: I1206 14:23:45.545397 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-62dkd_0514bb3a-f280-4208-9dff-ca5549dff0ca/registry-server/0.log" Dec 06 14:23:45 crc kubenswrapper[4706]: I1206 14:23:45.631063 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hl9mm_8278b19e-f7eb-43ee-8d01-473f09b98a95/extract-utilities/0.log" Dec 06 14:23:45 crc kubenswrapper[4706]: I1206 14:23:45.802955 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hl9mm_8278b19e-f7eb-43ee-8d01-473f09b98a95/extract-content/0.log" Dec 06 14:23:45 crc kubenswrapper[4706]: I1206 14:23:45.812091 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hl9mm_8278b19e-f7eb-43ee-8d01-473f09b98a95/extract-utilities/0.log" Dec 06 14:23:45 crc kubenswrapper[4706]: I1206 14:23:45.833205 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hl9mm_8278b19e-f7eb-43ee-8d01-473f09b98a95/extract-content/0.log" Dec 06 14:23:46 crc kubenswrapper[4706]: I1206 14:23:46.013669 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hl9mm_8278b19e-f7eb-43ee-8d01-473f09b98a95/extract-utilities/0.log" Dec 06 14:23:46 crc kubenswrapper[4706]: I1206 14:23:46.015494 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hl9mm_8278b19e-f7eb-43ee-8d01-473f09b98a95/extract-content/0.log" Dec 06 14:23:46 crc kubenswrapper[4706]: I1206 14:23:46.308819 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hl9mm_8278b19e-f7eb-43ee-8d01-473f09b98a95/registry-server/0.log" Dec 06 14:23:47 crc kubenswrapper[4706]: I1206 14:23:47.514576 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:23:47 crc kubenswrapper[4706]: E1206 14:23:47.514953 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:24:01 crc kubenswrapper[4706]: I1206 14:24:01.511526 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:24:01 crc kubenswrapper[4706]: E1206 14:24:01.512121 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:24:05 crc kubenswrapper[4706]: I1206 14:24:05.288455 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-59v8t"] Dec 06 14:24:05 crc kubenswrapper[4706]: E1206 14:24:05.289895 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca8afca-8dd4-475a-8f47-d492f2940eee" containerName="extract-content" Dec 06 14:24:05 crc kubenswrapper[4706]: I1206 14:24:05.289919 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca8afca-8dd4-475a-8f47-d492f2940eee" containerName="extract-content" Dec 06 14:24:05 crc kubenswrapper[4706]: E1206 14:24:05.289953 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca8afca-8dd4-475a-8f47-d492f2940eee" containerName="extract-utilities" Dec 06 14:24:05 crc kubenswrapper[4706]: I1206 14:24:05.289964 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca8afca-8dd4-475a-8f47-d492f2940eee" containerName="extract-utilities" Dec 06 14:24:05 crc kubenswrapper[4706]: E1206 14:24:05.289998 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca8afca-8dd4-475a-8f47-d492f2940eee" containerName="registry-server" Dec 06 14:24:05 crc kubenswrapper[4706]: I1206 14:24:05.290058 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca8afca-8dd4-475a-8f47-d492f2940eee" containerName="registry-server" Dec 06 14:24:05 crc kubenswrapper[4706]: I1206 14:24:05.293634 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ca8afca-8dd4-475a-8f47-d492f2940eee" containerName="registry-server" Dec 06 14:24:05 crc kubenswrapper[4706]: I1206 14:24:05.295531 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-59v8t" Dec 06 14:24:05 crc kubenswrapper[4706]: I1206 14:24:05.312241 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-59v8t"] Dec 06 14:24:05 crc kubenswrapper[4706]: I1206 14:24:05.346201 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8069e5cc-225d-469a-95a8-026470b7cf67-catalog-content\") pod \"community-operators-59v8t\" (UID: \"8069e5cc-225d-469a-95a8-026470b7cf67\") " pod="openshift-marketplace/community-operators-59v8t" Dec 06 14:24:05 crc kubenswrapper[4706]: I1206 14:24:05.346299 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp2ds\" (UniqueName: \"kubernetes.io/projected/8069e5cc-225d-469a-95a8-026470b7cf67-kube-api-access-rp2ds\") pod \"community-operators-59v8t\" (UID: \"8069e5cc-225d-469a-95a8-026470b7cf67\") " pod="openshift-marketplace/community-operators-59v8t" Dec 06 14:24:05 crc kubenswrapper[4706]: I1206 14:24:05.346355 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8069e5cc-225d-469a-95a8-026470b7cf67-utilities\") pod \"community-operators-59v8t\" (UID: \"8069e5cc-225d-469a-95a8-026470b7cf67\") " pod="openshift-marketplace/community-operators-59v8t" Dec 06 14:24:05 crc kubenswrapper[4706]: I1206 14:24:05.447130 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8069e5cc-225d-469a-95a8-026470b7cf67-catalog-content\") pod \"community-operators-59v8t\" (UID: \"8069e5cc-225d-469a-95a8-026470b7cf67\") " pod="openshift-marketplace/community-operators-59v8t" Dec 06 14:24:05 crc kubenswrapper[4706]: I1206 14:24:05.447221 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8069e5cc-225d-469a-95a8-026470b7cf67-utilities\") pod \"community-operators-59v8t\" (UID: \"8069e5cc-225d-469a-95a8-026470b7cf67\") " pod="openshift-marketplace/community-operators-59v8t" Dec 06 14:24:05 crc kubenswrapper[4706]: I1206 14:24:05.447249 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp2ds\" (UniqueName: \"kubernetes.io/projected/8069e5cc-225d-469a-95a8-026470b7cf67-kube-api-access-rp2ds\") pod \"community-operators-59v8t\" (UID: \"8069e5cc-225d-469a-95a8-026470b7cf67\") " pod="openshift-marketplace/community-operators-59v8t" Dec 06 14:24:05 crc kubenswrapper[4706]: I1206 14:24:05.447739 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8069e5cc-225d-469a-95a8-026470b7cf67-catalog-content\") pod \"community-operators-59v8t\" (UID: \"8069e5cc-225d-469a-95a8-026470b7cf67\") " pod="openshift-marketplace/community-operators-59v8t" Dec 06 14:24:05 crc kubenswrapper[4706]: I1206 14:24:05.447781 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8069e5cc-225d-469a-95a8-026470b7cf67-utilities\") pod \"community-operators-59v8t\" (UID: \"8069e5cc-225d-469a-95a8-026470b7cf67\") " pod="openshift-marketplace/community-operators-59v8t" Dec 06 14:24:05 crc kubenswrapper[4706]: I1206 14:24:05.464783 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp2ds\" (UniqueName: \"kubernetes.io/projected/8069e5cc-225d-469a-95a8-026470b7cf67-kube-api-access-rp2ds\") pod \"community-operators-59v8t\" (UID: \"8069e5cc-225d-469a-95a8-026470b7cf67\") " pod="openshift-marketplace/community-operators-59v8t" Dec 06 14:24:05 crc kubenswrapper[4706]: I1206 14:24:05.624243 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-59v8t" Dec 06 14:24:06 crc kubenswrapper[4706]: I1206 14:24:06.116794 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-59v8t"] Dec 06 14:24:06 crc kubenswrapper[4706]: I1206 14:24:06.297980 4706 generic.go:334] "Generic (PLEG): container finished" podID="8069e5cc-225d-469a-95a8-026470b7cf67" containerID="349c1ff6b257fb7412ac3727ff6b5091cd62f782147903993ddecb3f38554d0c" exitCode=0 Dec 06 14:24:06 crc kubenswrapper[4706]: I1206 14:24:06.298096 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-59v8t" event={"ID":"8069e5cc-225d-469a-95a8-026470b7cf67","Type":"ContainerDied","Data":"349c1ff6b257fb7412ac3727ff6b5091cd62f782147903993ddecb3f38554d0c"} Dec 06 14:24:06 crc kubenswrapper[4706]: I1206 14:24:06.298243 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-59v8t" event={"ID":"8069e5cc-225d-469a-95a8-026470b7cf67","Type":"ContainerStarted","Data":"ca4f221914f6de94f2d930be07a8c42ce16b188e745a2da893276d432fe355f7"} Dec 06 14:24:07 crc kubenswrapper[4706]: I1206 14:24:07.305072 4706 generic.go:334] "Generic (PLEG): container finished" podID="8069e5cc-225d-469a-95a8-026470b7cf67" containerID="51be36c911a8681a9c09de1767e81cf7da35c478d67e89069aef50522e94619a" exitCode=0 Dec 06 14:24:07 crc kubenswrapper[4706]: I1206 14:24:07.305137 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-59v8t" event={"ID":"8069e5cc-225d-469a-95a8-026470b7cf67","Type":"ContainerDied","Data":"51be36c911a8681a9c09de1767e81cf7da35c478d67e89069aef50522e94619a"} Dec 06 14:24:08 crc kubenswrapper[4706]: I1206 14:24:08.317768 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-59v8t" event={"ID":"8069e5cc-225d-469a-95a8-026470b7cf67","Type":"ContainerStarted","Data":"9ec7772db1ebee24f293068edead1d16dddefb20a2ea616f6e6ebc352a181cea"} Dec 06 14:24:08 crc kubenswrapper[4706]: I1206 14:24:08.355885 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-59v8t" podStartSLOduration=1.965428411 podStartE2EDuration="3.355852306s" podCreationTimestamp="2025-12-06 14:24:05 +0000 UTC" firstStartedPulling="2025-12-06 14:24:06.300386613 +0000 UTC m=+1689.136179470" lastFinishedPulling="2025-12-06 14:24:07.690810478 +0000 UTC m=+1690.526603365" observedRunningTime="2025-12-06 14:24:08.348217334 +0000 UTC m=+1691.184010231" watchObservedRunningTime="2025-12-06 14:24:08.355852306 +0000 UTC m=+1691.191645193" Dec 06 14:24:15 crc kubenswrapper[4706]: I1206 14:24:15.510905 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:24:15 crc kubenswrapper[4706]: E1206 14:24:15.511729 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:24:15 crc kubenswrapper[4706]: I1206 14:24:15.625926 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-59v8t" Dec 06 14:24:15 crc kubenswrapper[4706]: I1206 14:24:15.626407 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-59v8t" Dec 06 14:24:15 crc kubenswrapper[4706]: I1206 14:24:15.674440 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-59v8t" Dec 06 14:24:16 crc kubenswrapper[4706]: I1206 14:24:16.430469 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-59v8t" Dec 06 14:24:16 crc kubenswrapper[4706]: I1206 14:24:16.489726 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-59v8t"] Dec 06 14:24:18 crc kubenswrapper[4706]: I1206 14:24:18.398976 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-59v8t" podUID="8069e5cc-225d-469a-95a8-026470b7cf67" containerName="registry-server" containerID="cri-o://9ec7772db1ebee24f293068edead1d16dddefb20a2ea616f6e6ebc352a181cea" gracePeriod=2 Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.377245 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-59v8t" Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.407957 4706 generic.go:334] "Generic (PLEG): container finished" podID="8069e5cc-225d-469a-95a8-026470b7cf67" containerID="9ec7772db1ebee24f293068edead1d16dddefb20a2ea616f6e6ebc352a181cea" exitCode=0 Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.408000 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-59v8t" event={"ID":"8069e5cc-225d-469a-95a8-026470b7cf67","Type":"ContainerDied","Data":"9ec7772db1ebee24f293068edead1d16dddefb20a2ea616f6e6ebc352a181cea"} Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.408026 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-59v8t" event={"ID":"8069e5cc-225d-469a-95a8-026470b7cf67","Type":"ContainerDied","Data":"ca4f221914f6de94f2d930be07a8c42ce16b188e745a2da893276d432fe355f7"} Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.408044 4706 scope.go:117] "RemoveContainer" containerID="9ec7772db1ebee24f293068edead1d16dddefb20a2ea616f6e6ebc352a181cea" Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.408108 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-59v8t" Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.427409 4706 scope.go:117] "RemoveContainer" containerID="51be36c911a8681a9c09de1767e81cf7da35c478d67e89069aef50522e94619a" Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.448861 4706 scope.go:117] "RemoveContainer" containerID="349c1ff6b257fb7412ac3727ff6b5091cd62f782147903993ddecb3f38554d0c" Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.482939 4706 scope.go:117] "RemoveContainer" containerID="9ec7772db1ebee24f293068edead1d16dddefb20a2ea616f6e6ebc352a181cea" Dec 06 14:24:19 crc kubenswrapper[4706]: E1206 14:24:19.483294 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ec7772db1ebee24f293068edead1d16dddefb20a2ea616f6e6ebc352a181cea\": container with ID starting with 9ec7772db1ebee24f293068edead1d16dddefb20a2ea616f6e6ebc352a181cea not found: ID does not exist" containerID="9ec7772db1ebee24f293068edead1d16dddefb20a2ea616f6e6ebc352a181cea" Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.483346 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ec7772db1ebee24f293068edead1d16dddefb20a2ea616f6e6ebc352a181cea"} err="failed to get container status \"9ec7772db1ebee24f293068edead1d16dddefb20a2ea616f6e6ebc352a181cea\": rpc error: code = NotFound desc = could not find container \"9ec7772db1ebee24f293068edead1d16dddefb20a2ea616f6e6ebc352a181cea\": container with ID starting with 9ec7772db1ebee24f293068edead1d16dddefb20a2ea616f6e6ebc352a181cea not found: ID does not exist" Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.483373 4706 scope.go:117] "RemoveContainer" containerID="51be36c911a8681a9c09de1767e81cf7da35c478d67e89069aef50522e94619a" Dec 06 14:24:19 crc kubenswrapper[4706]: E1206 14:24:19.483713 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51be36c911a8681a9c09de1767e81cf7da35c478d67e89069aef50522e94619a\": container with ID starting with 51be36c911a8681a9c09de1767e81cf7da35c478d67e89069aef50522e94619a not found: ID does not exist" containerID="51be36c911a8681a9c09de1767e81cf7da35c478d67e89069aef50522e94619a" Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.483776 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51be36c911a8681a9c09de1767e81cf7da35c478d67e89069aef50522e94619a"} err="failed to get container status \"51be36c911a8681a9c09de1767e81cf7da35c478d67e89069aef50522e94619a\": rpc error: code = NotFound desc = could not find container \"51be36c911a8681a9c09de1767e81cf7da35c478d67e89069aef50522e94619a\": container with ID starting with 51be36c911a8681a9c09de1767e81cf7da35c478d67e89069aef50522e94619a not found: ID does not exist" Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.483819 4706 scope.go:117] "RemoveContainer" containerID="349c1ff6b257fb7412ac3727ff6b5091cd62f782147903993ddecb3f38554d0c" Dec 06 14:24:19 crc kubenswrapper[4706]: E1206 14:24:19.484241 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"349c1ff6b257fb7412ac3727ff6b5091cd62f782147903993ddecb3f38554d0c\": container with ID starting with 349c1ff6b257fb7412ac3727ff6b5091cd62f782147903993ddecb3f38554d0c not found: ID does not exist" containerID="349c1ff6b257fb7412ac3727ff6b5091cd62f782147903993ddecb3f38554d0c" Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.484279 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"349c1ff6b257fb7412ac3727ff6b5091cd62f782147903993ddecb3f38554d0c"} err="failed to get container status \"349c1ff6b257fb7412ac3727ff6b5091cd62f782147903993ddecb3f38554d0c\": rpc error: code = NotFound desc = could not find container \"349c1ff6b257fb7412ac3727ff6b5091cd62f782147903993ddecb3f38554d0c\": container with ID starting with 349c1ff6b257fb7412ac3727ff6b5091cd62f782147903993ddecb3f38554d0c not found: ID does not exist" Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.545036 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8069e5cc-225d-469a-95a8-026470b7cf67-utilities\") pod \"8069e5cc-225d-469a-95a8-026470b7cf67\" (UID: \"8069e5cc-225d-469a-95a8-026470b7cf67\") " Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.546342 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8069e5cc-225d-469a-95a8-026470b7cf67-utilities" (OuterVolumeSpecName: "utilities") pod "8069e5cc-225d-469a-95a8-026470b7cf67" (UID: "8069e5cc-225d-469a-95a8-026470b7cf67"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.546922 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8069e5cc-225d-469a-95a8-026470b7cf67-catalog-content\") pod \"8069e5cc-225d-469a-95a8-026470b7cf67\" (UID: \"8069e5cc-225d-469a-95a8-026470b7cf67\") " Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.547146 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rp2ds\" (UniqueName: \"kubernetes.io/projected/8069e5cc-225d-469a-95a8-026470b7cf67-kube-api-access-rp2ds\") pod \"8069e5cc-225d-469a-95a8-026470b7cf67\" (UID: \"8069e5cc-225d-469a-95a8-026470b7cf67\") " Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.549251 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8069e5cc-225d-469a-95a8-026470b7cf67-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.554110 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8069e5cc-225d-469a-95a8-026470b7cf67-kube-api-access-rp2ds" (OuterVolumeSpecName: "kube-api-access-rp2ds") pod "8069e5cc-225d-469a-95a8-026470b7cf67" (UID: "8069e5cc-225d-469a-95a8-026470b7cf67"). InnerVolumeSpecName "kube-api-access-rp2ds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.604081 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8069e5cc-225d-469a-95a8-026470b7cf67-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8069e5cc-225d-469a-95a8-026470b7cf67" (UID: "8069e5cc-225d-469a-95a8-026470b7cf67"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.650786 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8069e5cc-225d-469a-95a8-026470b7cf67-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.650823 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rp2ds\" (UniqueName: \"kubernetes.io/projected/8069e5cc-225d-469a-95a8-026470b7cf67-kube-api-access-rp2ds\") on node \"crc\" DevicePath \"\"" Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.768493 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-59v8t"] Dec 06 14:24:19 crc kubenswrapper[4706]: I1206 14:24:19.774736 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-59v8t"] Dec 06 14:24:21 crc kubenswrapper[4706]: I1206 14:24:21.523676 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8069e5cc-225d-469a-95a8-026470b7cf67" path="/var/lib/kubelet/pods/8069e5cc-225d-469a-95a8-026470b7cf67/volumes" Dec 06 14:24:26 crc kubenswrapper[4706]: I1206 14:24:26.514117 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:24:26 crc kubenswrapper[4706]: E1206 14:24:26.515440 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:24:39 crc kubenswrapper[4706]: I1206 14:24:39.511625 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:24:39 crc kubenswrapper[4706]: E1206 14:24:39.513555 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:24:49 crc kubenswrapper[4706]: I1206 14:24:49.619356 4706 generic.go:334] "Generic (PLEG): container finished" podID="b893ca08-355d-41e9-a9a1-18cceff55185" containerID="cb54c9e5cec94e1c193edea9567f6bd3ab1ea10c516be0d27a12e17b924eeabf" exitCode=0 Dec 06 14:24:49 crc kubenswrapper[4706]: I1206 14:24:49.619461 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zwm4d/must-gather-skpfz" event={"ID":"b893ca08-355d-41e9-a9a1-18cceff55185","Type":"ContainerDied","Data":"cb54c9e5cec94e1c193edea9567f6bd3ab1ea10c516be0d27a12e17b924eeabf"} Dec 06 14:24:49 crc kubenswrapper[4706]: I1206 14:24:49.620441 4706 scope.go:117] "RemoveContainer" containerID="cb54c9e5cec94e1c193edea9567f6bd3ab1ea10c516be0d27a12e17b924eeabf" Dec 06 14:24:50 crc kubenswrapper[4706]: I1206 14:24:50.597791 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zwm4d_must-gather-skpfz_b893ca08-355d-41e9-a9a1-18cceff55185/gather/0.log" Dec 06 14:24:53 crc kubenswrapper[4706]: I1206 14:24:53.511922 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:24:53 crc kubenswrapper[4706]: E1206 14:24:53.512445 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.019332 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zwm4d/must-gather-skpfz"] Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.020053 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-zwm4d/must-gather-skpfz" podUID="b893ca08-355d-41e9-a9a1-18cceff55185" containerName="copy" containerID="cri-o://4c9d04651e818fa57bf3bcc11e84f04c4daf2a44a8bfafbccf2c3c9f93b55795" gracePeriod=2 Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.023846 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zwm4d/must-gather-skpfz"] Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.385491 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zwm4d_must-gather-skpfz_b893ca08-355d-41e9-a9a1-18cceff55185/copy/0.log" Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.386397 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zwm4d/must-gather-skpfz" Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.577827 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffjd2\" (UniqueName: \"kubernetes.io/projected/b893ca08-355d-41e9-a9a1-18cceff55185-kube-api-access-ffjd2\") pod \"b893ca08-355d-41e9-a9a1-18cceff55185\" (UID: \"b893ca08-355d-41e9-a9a1-18cceff55185\") " Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.577900 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b893ca08-355d-41e9-a9a1-18cceff55185-must-gather-output\") pod \"b893ca08-355d-41e9-a9a1-18cceff55185\" (UID: \"b893ca08-355d-41e9-a9a1-18cceff55185\") " Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.584410 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b893ca08-355d-41e9-a9a1-18cceff55185-kube-api-access-ffjd2" (OuterVolumeSpecName: "kube-api-access-ffjd2") pod "b893ca08-355d-41e9-a9a1-18cceff55185" (UID: "b893ca08-355d-41e9-a9a1-18cceff55185"). InnerVolumeSpecName "kube-api-access-ffjd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.631938 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b893ca08-355d-41e9-a9a1-18cceff55185-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b893ca08-355d-41e9-a9a1-18cceff55185" (UID: "b893ca08-355d-41e9-a9a1-18cceff55185"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.680157 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffjd2\" (UniqueName: \"kubernetes.io/projected/b893ca08-355d-41e9-a9a1-18cceff55185-kube-api-access-ffjd2\") on node \"crc\" DevicePath \"\"" Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.680268 4706 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b893ca08-355d-41e9-a9a1-18cceff55185-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.701538 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zwm4d_must-gather-skpfz_b893ca08-355d-41e9-a9a1-18cceff55185/copy/0.log" Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.702152 4706 generic.go:334] "Generic (PLEG): container finished" podID="b893ca08-355d-41e9-a9a1-18cceff55185" containerID="4c9d04651e818fa57bf3bcc11e84f04c4daf2a44a8bfafbccf2c3c9f93b55795" exitCode=143 Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.702235 4706 scope.go:117] "RemoveContainer" containerID="4c9d04651e818fa57bf3bcc11e84f04c4daf2a44a8bfafbccf2c3c9f93b55795" Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.702273 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zwm4d/must-gather-skpfz" Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.725277 4706 scope.go:117] "RemoveContainer" containerID="cb54c9e5cec94e1c193edea9567f6bd3ab1ea10c516be0d27a12e17b924eeabf" Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.786012 4706 scope.go:117] "RemoveContainer" containerID="4c9d04651e818fa57bf3bcc11e84f04c4daf2a44a8bfafbccf2c3c9f93b55795" Dec 06 14:25:00 crc kubenswrapper[4706]: E1206 14:25:00.786714 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c9d04651e818fa57bf3bcc11e84f04c4daf2a44a8bfafbccf2c3c9f93b55795\": container with ID starting with 4c9d04651e818fa57bf3bcc11e84f04c4daf2a44a8bfafbccf2c3c9f93b55795 not found: ID does not exist" containerID="4c9d04651e818fa57bf3bcc11e84f04c4daf2a44a8bfafbccf2c3c9f93b55795" Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.786789 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c9d04651e818fa57bf3bcc11e84f04c4daf2a44a8bfafbccf2c3c9f93b55795"} err="failed to get container status \"4c9d04651e818fa57bf3bcc11e84f04c4daf2a44a8bfafbccf2c3c9f93b55795\": rpc error: code = NotFound desc = could not find container \"4c9d04651e818fa57bf3bcc11e84f04c4daf2a44a8bfafbccf2c3c9f93b55795\": container with ID starting with 4c9d04651e818fa57bf3bcc11e84f04c4daf2a44a8bfafbccf2c3c9f93b55795 not found: ID does not exist" Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.786865 4706 scope.go:117] "RemoveContainer" containerID="cb54c9e5cec94e1c193edea9567f6bd3ab1ea10c516be0d27a12e17b924eeabf" Dec 06 14:25:00 crc kubenswrapper[4706]: E1206 14:25:00.787265 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb54c9e5cec94e1c193edea9567f6bd3ab1ea10c516be0d27a12e17b924eeabf\": container with ID starting with cb54c9e5cec94e1c193edea9567f6bd3ab1ea10c516be0d27a12e17b924eeabf not found: ID does not exist" containerID="cb54c9e5cec94e1c193edea9567f6bd3ab1ea10c516be0d27a12e17b924eeabf" Dec 06 14:25:00 crc kubenswrapper[4706]: I1206 14:25:00.787299 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb54c9e5cec94e1c193edea9567f6bd3ab1ea10c516be0d27a12e17b924eeabf"} err="failed to get container status \"cb54c9e5cec94e1c193edea9567f6bd3ab1ea10c516be0d27a12e17b924eeabf\": rpc error: code = NotFound desc = could not find container \"cb54c9e5cec94e1c193edea9567f6bd3ab1ea10c516be0d27a12e17b924eeabf\": container with ID starting with cb54c9e5cec94e1c193edea9567f6bd3ab1ea10c516be0d27a12e17b924eeabf not found: ID does not exist" Dec 06 14:25:01 crc kubenswrapper[4706]: I1206 14:25:01.522945 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b893ca08-355d-41e9-a9a1-18cceff55185" path="/var/lib/kubelet/pods/b893ca08-355d-41e9-a9a1-18cceff55185/volumes" Dec 06 14:25:07 crc kubenswrapper[4706]: I1206 14:25:07.516795 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:25:07 crc kubenswrapper[4706]: E1206 14:25:07.517596 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:25:22 crc kubenswrapper[4706]: I1206 14:25:22.510915 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:25:22 crc kubenswrapper[4706]: E1206 14:25:22.511572 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:25:37 crc kubenswrapper[4706]: I1206 14:25:37.517464 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:25:37 crc kubenswrapper[4706]: E1206 14:25:37.518165 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:25:52 crc kubenswrapper[4706]: I1206 14:25:52.511671 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:25:52 crc kubenswrapper[4706]: E1206 14:25:52.512910 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:26:06 crc kubenswrapper[4706]: I1206 14:26:06.511816 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:26:06 crc kubenswrapper[4706]: E1206 14:26:06.512903 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:26:18 crc kubenswrapper[4706]: I1206 14:26:18.511788 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:26:18 crc kubenswrapper[4706]: E1206 14:26:18.512838 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:26:32 crc kubenswrapper[4706]: I1206 14:26:32.511724 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:26:32 crc kubenswrapper[4706]: E1206 14:26:32.512571 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:26:46 crc kubenswrapper[4706]: I1206 14:26:46.511551 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:26:46 crc kubenswrapper[4706]: E1206 14:26:46.512486 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:26:58 crc kubenswrapper[4706]: I1206 14:26:58.511382 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:26:58 crc kubenswrapper[4706]: E1206 14:26:58.512176 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:27:10 crc kubenswrapper[4706]: I1206 14:27:10.512603 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:27:10 crc kubenswrapper[4706]: E1206 14:27:10.513464 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-67v56_openshift-machine-config-operator(904f5a16-0c23-4a2a-a945-59b1142cb4e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-67v56" podUID="904f5a16-0c23-4a2a-a945-59b1142cb4e1" Dec 06 14:27:25 crc kubenswrapper[4706]: I1206 14:27:25.511523 4706 scope.go:117] "RemoveContainer" containerID="2f0dff88651a7cee8ccffe471aea5a0801f3e1ab9bf2c8fe5ab9fd3957c6589f" Dec 06 14:27:25 crc kubenswrapper[4706]: I1206 14:27:25.839855 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-67v56" event={"ID":"904f5a16-0c23-4a2a-a945-59b1142cb4e1","Type":"ContainerStarted","Data":"87cc71330cce178689451ef6d8e43358740df6dd9d8ed8ade0205348f513cef2"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515115036574024454 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015115036575017372 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015115032475016510 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015115032476015461 5ustar corecore